2018-11-06 05:39:48 +00:00
|
|
|
package docker
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
"os"
|
|
|
|
"runtime"
|
|
|
|
"strings"
|
|
|
|
"sync"
|
|
|
|
"syscall"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/armon/circbuf"
|
|
|
|
docker "github.com/fsouza/go-dockerclient"
|
|
|
|
hclog "github.com/hashicorp/go-hclog"
|
|
|
|
plugin "github.com/hashicorp/go-plugin"
|
2018-11-28 16:29:57 +00:00
|
|
|
cstructs "github.com/hashicorp/nomad/client/structs"
|
2018-11-06 05:39:48 +00:00
|
|
|
"github.com/hashicorp/nomad/drivers/docker/docklog"
|
|
|
|
"github.com/hashicorp/nomad/helper/stats"
|
|
|
|
"github.com/hashicorp/nomad/plugins/drivers"
|
2018-11-19 18:15:15 +00:00
|
|
|
"github.com/hashicorp/nomad/plugins/shared"
|
2018-11-06 05:39:48 +00:00
|
|
|
"golang.org/x/net/context"
|
|
|
|
)
|
|
|
|
|
|
|
|
type taskHandle struct {
|
|
|
|
client *docker.Client
|
|
|
|
waitClient *docker.Client
|
|
|
|
logger hclog.Logger
|
|
|
|
dlogger docklog.DockerLogger
|
|
|
|
dloggerPluginClient *plugin.Client
|
|
|
|
task *drivers.TaskConfig
|
2018-11-20 02:41:25 +00:00
|
|
|
containerID string
|
|
|
|
containerImage string
|
2018-11-06 05:39:48 +00:00
|
|
|
resourceUsageLock sync.RWMutex
|
2018-11-28 16:29:57 +00:00
|
|
|
resourceUsage *cstructs.TaskResourceUsage
|
2018-11-06 05:39:48 +00:00
|
|
|
doneCh chan bool
|
|
|
|
waitCh chan struct{}
|
|
|
|
removeContainerOnExit bool
|
2018-11-28 16:29:57 +00:00
|
|
|
net *cstructs.DriverNetwork
|
2018-11-06 05:39:48 +00:00
|
|
|
|
2018-11-21 01:41:32 +00:00
|
|
|
exitResult *drivers.ExitResult
|
|
|
|
exitResultLock sync.Mutex
|
|
|
|
}
|
|
|
|
|
|
|
|
func (h *taskHandle) ExitResult() *drivers.ExitResult {
|
|
|
|
h.exitResultLock.Lock()
|
|
|
|
defer h.exitResultLock.Unlock()
|
|
|
|
return h.exitResult.Copy()
|
2018-11-06 05:39:48 +00:00
|
|
|
}
|
|
|
|
|
2018-11-12 18:51:53 +00:00
|
|
|
type taskHandleState struct {
|
2018-11-16 18:52:54 +00:00
|
|
|
// ReattachConfig for the docker logger plugin
|
2018-11-19 18:15:15 +00:00
|
|
|
ReattachConfig *shared.ReattachConfig
|
2018-11-16 18:52:54 +00:00
|
|
|
|
|
|
|
ContainerID string
|
2018-11-28 16:29:57 +00:00
|
|
|
DriverNetwork *cstructs.DriverNetwork
|
2018-11-12 18:51:53 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (h *taskHandle) buildState() *taskHandleState {
|
|
|
|
return &taskHandleState{
|
2018-11-19 18:15:15 +00:00
|
|
|
ReattachConfig: shared.ReattachConfigFromGoPlugin(h.dloggerPluginClient.ReattachConfig()),
|
2018-11-20 02:41:25 +00:00
|
|
|
ContainerID: h.containerID,
|
2018-11-12 18:51:53 +00:00
|
|
|
DriverNetwork: h.net,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-11-06 05:39:48 +00:00
|
|
|
func (h *taskHandle) Exec(ctx context.Context, cmd string, args []string) (*drivers.ExecTaskResult, error) {
|
|
|
|
fullCmd := make([]string, len(args)+1)
|
|
|
|
fullCmd[0] = cmd
|
|
|
|
copy(fullCmd[1:], args)
|
|
|
|
createExecOpts := docker.CreateExecOptions{
|
|
|
|
AttachStdin: false,
|
|
|
|
AttachStdout: true,
|
|
|
|
AttachStderr: true,
|
|
|
|
Tty: false,
|
|
|
|
Cmd: fullCmd,
|
2018-11-20 02:41:25 +00:00
|
|
|
Container: h.containerID,
|
2018-11-06 05:39:48 +00:00
|
|
|
Context: ctx,
|
|
|
|
}
|
|
|
|
exec, err := h.client.CreateExec(createExecOpts)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
execResult := &drivers.ExecTaskResult{ExitResult: &drivers.ExitResult{}}
|
2018-11-30 09:56:45 +00:00
|
|
|
stdout, _ := circbuf.NewBuffer(int64(cstructs.CheckBufSize))
|
|
|
|
stderr, _ := circbuf.NewBuffer(int64(cstructs.CheckBufSize))
|
2018-11-06 05:39:48 +00:00
|
|
|
startOpts := docker.StartExecOptions{
|
|
|
|
Detach: false,
|
|
|
|
Tty: false,
|
|
|
|
OutputStream: stdout,
|
|
|
|
ErrorStream: stderr,
|
|
|
|
Context: ctx,
|
|
|
|
}
|
|
|
|
if err := client.StartExec(exec.ID, startOpts); err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
execResult.Stdout = stdout.Bytes()
|
|
|
|
execResult.Stderr = stderr.Bytes()
|
|
|
|
res, err := client.InspectExec(exec.ID)
|
|
|
|
if err != nil {
|
|
|
|
return execResult, err
|
|
|
|
}
|
|
|
|
|
|
|
|
execResult.ExitResult.ExitCode = res.ExitCode
|
|
|
|
return execResult, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (h *taskHandle) Signal(s os.Signal) error {
|
|
|
|
// Convert types
|
|
|
|
sysSig, ok := s.(syscall.Signal)
|
|
|
|
if !ok {
|
|
|
|
return fmt.Errorf("Failed to determine signal number")
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO When we expose signals we will need a mapping layer that converts
|
|
|
|
// MacOS signals to the correct signal number for docker. Or we change the
|
|
|
|
// interface to take a signal string and leave it up to driver to map?
|
|
|
|
|
|
|
|
dockerSignal := docker.Signal(sysSig)
|
|
|
|
opts := docker.KillContainerOptions{
|
2018-11-20 02:41:25 +00:00
|
|
|
ID: h.containerID,
|
2018-11-06 05:39:48 +00:00
|
|
|
Signal: dockerSignal,
|
|
|
|
}
|
|
|
|
return h.client.KillContainer(opts)
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
// Kill is used to terminate the task.
|
|
|
|
func (h *taskHandle) Kill(killTimeout time.Duration, signal os.Signal) error {
|
|
|
|
// Only send signal if killTimeout is set, otherwise stop container
|
|
|
|
if killTimeout > 0 {
|
|
|
|
if err := h.Signal(signal); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case <-h.waitCh:
|
|
|
|
return nil
|
|
|
|
case <-time.After(killTimeout):
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Stop the container
|
2018-11-20 02:41:25 +00:00
|
|
|
err := h.client.StopContainer(h.containerID, 0)
|
2018-11-06 05:39:48 +00:00
|
|
|
if err != nil {
|
|
|
|
|
|
|
|
// Container has already been removed.
|
|
|
|
if strings.Contains(err.Error(), NoSuchContainerError) {
|
|
|
|
h.logger.Debug("attempted to stop nonexistent container")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
h.logger.Error("failed to stop container", "error", err)
|
2018-11-20 02:41:25 +00:00
|
|
|
return fmt.Errorf("Failed to stop container %s: %s", h.containerID, err)
|
2018-11-06 05:39:48 +00:00
|
|
|
}
|
|
|
|
h.logger.Info("stopped container")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2018-11-28 16:29:57 +00:00
|
|
|
func (h *taskHandle) Stats() (*cstructs.TaskResourceUsage, error) {
|
2018-11-06 05:39:48 +00:00
|
|
|
h.resourceUsageLock.RLock()
|
|
|
|
defer h.resourceUsageLock.RUnlock()
|
|
|
|
var err error
|
|
|
|
if h.resourceUsage == nil {
|
|
|
|
err = fmt.Errorf("stats collection hasn't started yet")
|
|
|
|
}
|
|
|
|
return h.resourceUsage, err
|
|
|
|
}
|
|
|
|
|
|
|
|
func (h *taskHandle) run() {
|
2018-11-20 02:41:25 +00:00
|
|
|
exitCode, werr := h.waitClient.WaitContainer(h.containerID)
|
2018-11-06 05:39:48 +00:00
|
|
|
if werr != nil {
|
|
|
|
h.logger.Error("failed to wait for container; already terminated")
|
|
|
|
}
|
|
|
|
|
|
|
|
if exitCode != 0 {
|
|
|
|
werr = fmt.Errorf("Docker container exited with non-zero exit code: %d", exitCode)
|
|
|
|
}
|
|
|
|
|
2018-11-20 02:41:25 +00:00
|
|
|
container, ierr := h.waitClient.InspectContainer(h.containerID)
|
2018-11-06 05:39:48 +00:00
|
|
|
oom := false
|
|
|
|
if ierr != nil {
|
|
|
|
h.logger.Error("failed to inspect container", "error", ierr)
|
|
|
|
} else if container.State.OOMKilled {
|
|
|
|
oom = true
|
|
|
|
werr = fmt.Errorf("OOM Killed")
|
|
|
|
}
|
|
|
|
|
|
|
|
close(h.doneCh)
|
|
|
|
|
|
|
|
// Shutdown the syslog collector
|
|
|
|
|
|
|
|
// Stop the container just incase the docker daemon's wait returned
|
|
|
|
// incorrectly
|
2018-11-20 02:41:25 +00:00
|
|
|
if err := h.client.StopContainer(h.containerID, 0); err != nil {
|
2018-11-06 05:39:48 +00:00
|
|
|
_, noSuchContainer := err.(*docker.NoSuchContainer)
|
|
|
|
_, containerNotRunning := err.(*docker.ContainerNotRunning)
|
|
|
|
if !containerNotRunning && !noSuchContainer {
|
|
|
|
h.logger.Error("error stopping container", "error", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Remove the container
|
|
|
|
if h.removeContainerOnExit == true {
|
2018-11-20 02:41:25 +00:00
|
|
|
if err := h.client.RemoveContainer(docker.RemoveContainerOptions{ID: h.containerID, RemoveVolumes: true, Force: true}); err != nil {
|
2018-11-06 05:39:48 +00:00
|
|
|
h.logger.Error("error removing container", "error", err)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
h.logger.Debug("not removing container due to config")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set the result
|
2018-11-21 01:41:32 +00:00
|
|
|
h.exitResultLock.Lock()
|
2018-11-06 05:39:48 +00:00
|
|
|
h.exitResult = &drivers.ExitResult{
|
|
|
|
ExitCode: exitCode,
|
|
|
|
Signal: 0,
|
|
|
|
OOMKilled: oom,
|
2018-11-14 11:20:35 +00:00
|
|
|
Err: werr,
|
2018-11-06 05:39:48 +00:00
|
|
|
}
|
2018-11-21 01:41:32 +00:00
|
|
|
h.exitResultLock.Unlock()
|
2018-11-06 05:39:48 +00:00
|
|
|
close(h.waitCh)
|
|
|
|
}
|
|
|
|
|
|
|
|
// collectStats starts collecting resource usage stats of a docker container
|
|
|
|
func (h *taskHandle) collectStats() {
|
|
|
|
|
|
|
|
statsCh := make(chan *docker.Stats)
|
2018-11-20 02:41:25 +00:00
|
|
|
statsOpts := docker.StatsOptions{ID: h.containerID, Done: h.doneCh, Stats: statsCh, Stream: true}
|
2018-11-06 05:39:48 +00:00
|
|
|
go func() {
|
|
|
|
//TODO handle Stats error
|
|
|
|
if err := h.waitClient.Stats(statsOpts); err != nil {
|
|
|
|
h.logger.Debug("error collecting stats from container", "error", err)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
numCores := runtime.NumCPU()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case s := <-statsCh:
|
|
|
|
if s != nil {
|
2018-11-28 16:29:57 +00:00
|
|
|
ms := &cstructs.MemoryStats{
|
2018-11-06 05:39:48 +00:00
|
|
|
RSS: s.MemoryStats.Stats.Rss,
|
|
|
|
Cache: s.MemoryStats.Stats.Cache,
|
|
|
|
Swap: s.MemoryStats.Stats.Swap,
|
|
|
|
MaxUsage: s.MemoryStats.MaxUsage,
|
|
|
|
Measured: DockerMeasuredMemStats,
|
|
|
|
}
|
|
|
|
|
2018-11-28 16:29:57 +00:00
|
|
|
cs := &cstructs.CpuStats{
|
2018-11-06 05:39:48 +00:00
|
|
|
ThrottledPeriods: s.CPUStats.ThrottlingData.ThrottledPeriods,
|
|
|
|
ThrottledTime: s.CPUStats.ThrottlingData.ThrottledTime,
|
|
|
|
Measured: DockerMeasuredCpuStats,
|
|
|
|
}
|
|
|
|
|
|
|
|
// Calculate percentage
|
|
|
|
cs.Percent = calculatePercent(
|
|
|
|
s.CPUStats.CPUUsage.TotalUsage, s.PreCPUStats.CPUUsage.TotalUsage,
|
|
|
|
s.CPUStats.SystemCPUUsage, s.PreCPUStats.SystemCPUUsage, numCores)
|
|
|
|
cs.SystemMode = calculatePercent(
|
|
|
|
s.CPUStats.CPUUsage.UsageInKernelmode, s.PreCPUStats.CPUUsage.UsageInKernelmode,
|
|
|
|
s.CPUStats.CPUUsage.TotalUsage, s.PreCPUStats.CPUUsage.TotalUsage, numCores)
|
|
|
|
cs.UserMode = calculatePercent(
|
|
|
|
s.CPUStats.CPUUsage.UsageInUsermode, s.PreCPUStats.CPUUsage.UsageInUsermode,
|
|
|
|
s.CPUStats.CPUUsage.TotalUsage, s.PreCPUStats.CPUUsage.TotalUsage, numCores)
|
|
|
|
cs.TotalTicks = (cs.Percent / 100) * stats.TotalTicksAvailable() / float64(numCores)
|
|
|
|
|
|
|
|
h.resourceUsageLock.Lock()
|
2018-11-28 16:29:57 +00:00
|
|
|
h.resourceUsage = &cstructs.TaskResourceUsage{
|
|
|
|
ResourceUsage: &cstructs.ResourceUsage{
|
2018-11-06 05:39:48 +00:00
|
|
|
MemoryStats: ms,
|
|
|
|
CpuStats: cs,
|
|
|
|
},
|
|
|
|
Timestamp: s.Read.UTC().UnixNano(),
|
|
|
|
}
|
|
|
|
h.resourceUsageLock.Unlock()
|
|
|
|
}
|
|
|
|
case <-h.doneCh:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func calculatePercent(newSample, oldSample, newTotal, oldTotal uint64, cores int) float64 {
|
|
|
|
numerator := newSample - oldSample
|
|
|
|
denom := newTotal - oldTotal
|
|
|
|
if numerator <= 0 || denom <= 0 {
|
|
|
|
return 0.0
|
|
|
|
}
|
|
|
|
|
|
|
|
return (float64(numerator) / float64(denom)) * float64(cores) * 100.0
|
|
|
|
}
|