package rkt import ( "bytes" "context" "fmt" "io/ioutil" "os" "path/filepath" "sync" "testing" "time" ctestutil "github.com/hashicorp/nomad/client/testutil" "github.com/hashicorp/nomad/helper/pluginutils/hclutils" "github.com/hashicorp/nomad/helper/testlog" "github.com/hashicorp/nomad/helper/testtask" "github.com/hashicorp/nomad/helper/uuid" "github.com/hashicorp/nomad/nomad/structs" basePlug "github.com/hashicorp/nomad/plugins/base" "github.com/hashicorp/nomad/plugins/drivers" dtestutil "github.com/hashicorp/nomad/plugins/drivers/testutils" "github.com/hashicorp/nomad/testutil" "github.com/stretchr/testify/require" "golang.org/x/sys/unix" ) var _ drivers.DriverPlugin = (*Driver)(nil) func TestRktVersionRegex(t *testing.T) { ctestutil.RktCompatible(t) t.Parallel() inputRkt := "rkt version 0.8.1" inputAppc := "appc version 1.2.0" expectedRkt := "0.8.1" expectedAppc := "1.2.0" rktMatches := reRktVersion.FindStringSubmatch(inputRkt) appcMatches := reAppcVersion.FindStringSubmatch(inputAppc) if rktMatches[1] != expectedRkt { fmt.Printf("Test failed; got %q; want %q\n", rktMatches[1], expectedRkt) } if appcMatches[1] != expectedAppc { fmt.Printf("Test failed; got %q; want %q\n", appcMatches[1], expectedAppc) } } // Tests setting driver config options func TestRktDriver_SetConfig(t *testing.T) { t.Parallel() require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) // Enable Volumes config := &Config{ VolumesEnabled: true, } var data []byte require.NoError(basePlug.MsgPackEncode(&data, config)) bconfig := &basePlug.Config{PluginConfig: data} require.NoError(harness.SetConfig(bconfig)) require.Exactly(config, d.(*Driver).config) config.VolumesEnabled = false data = []byte{} require.NoError(basePlug.MsgPackEncode(&data, config)) bconfig = &basePlug.Config{PluginConfig: data} require.NoError(harness.SetConfig(bconfig)) require.Exactly(config, d.(*Driver).config) } // Verifies using a trust prefix and passing dns servers and search domains // Also verifies sending sigterm correctly stops the driver instance func TestRktDriver_Start_Wait_Stop_DNS(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ TrustPrefix: "coreos.com/etcd", ImageName: "coreos.com/etcd:v2.0.4", Command: "/etcd", DNSServers: []string{"8.8.8.8", "8.8.4.4"}, DNSSearchDomains: []string{"example.com", "example.org", "example.net"}, Net: []string{"host"}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() handle, driverNet, err := harness.StartTask(task) require.NoError(err) require.Nil(driverNet) ch, err := harness.WaitTask(context.Background(), handle.Config.ID) require.NoError(err) require.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second)) go func() { harness.StopTask(task.ID, 2*time.Second, "SIGTERM") }() select { case result := <-ch: require.Equal(int(unix.SIGTERM), result.Signal) case <-time.After(10 * time.Second): require.Fail("timeout waiting for task to shutdown") } // Ensure that the task is marked as dead, but account // for WaitTask() closing channel before internal state is updated testutil.WaitForResult(func() (bool, error) { status, err := harness.InspectTask(task.ID) if err != nil { return false, fmt.Errorf("inspecting task failed: %v", err) } if status.State != drivers.TaskStateExited { return false, fmt.Errorf("task hasn't exited yet; status: %v", status.State) } return true, nil }, func(err error) { require.NoError(err) }) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies waiting on task to exit cleanly func TestRktDriver_Start_Wait_Stop(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ TrustPrefix: "coreos.com/etcd", ImageName: "coreos.com/etcd:v2.0.4", Command: "/etcd", Args: []string{"--version"}, Net: []string{"none"}, Debug: true, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) cleanup := harness.MkAllocDir(task, true) defer cleanup() handle, _, err := harness.StartTask(task) require.NoError(err) // Wait on the task, it should exit since we are only asking for etcd version here ch, err := harness.WaitTask(context.Background(), handle.Config.ID) require.NoError(err) result := <-ch require.Nil(result.Err) require.Zero(result.ExitCode) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies that skipping trust_prefix works func TestRktDriver_Start_Wait_Skip_Trust(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ ImageName: "coreos.com/etcd:v2.0.4", Command: "/etcd", Args: []string{"--version"}, Net: []string{"none"}, Debug: true, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() handle, _, err := harness.StartTask(task) require.NoError(err) // Wait on the task, it should exit since we are only asking for etcd version here ch, err := harness.WaitTask(context.Background(), handle.Config.ID) require.NoError(err) result := <-ch require.Nil(result.Err) require.Zero(result.ExitCode) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies that an invalid trust prefix returns expected error func TestRktDriver_InvalidTrustPrefix(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ TrustPrefix: "example.com/invalid", ImageName: "coreos.com/etcd:v2.0.4", Command: "/etcd", Args: []string{"--version"}, Net: []string{"none"}, Debug: true, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, _, err := harness.StartTask(task) require.Error(err) expectedErr := "Error running rkt trust" require.Contains(err.Error(), expectedErr) } // Verifies reattaching to a running container // This test manipulates the harness's internal state map // to remove the task and then reattaches to it func TestRktDriver_StartWaitRecoverWaitStop(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ ImageName: "coreos.com/etcd:v2.0.4", Command: "/etcd", } require.NoError(task.EncodeConcreteDriverConfig(&tc)) cleanup := harness.MkAllocDir(task, true) defer cleanup() handle, _, err := harness.StartTask(task) require.NoError(err) ch, err := harness.WaitTask(context.Background(), task.ID) require.NoError(err) var waitDone bool var wg sync.WaitGroup wg.Add(1) go func() { defer wg.Done() result := <-ch require.Error(result.Err) waitDone = true }() originalStatus, err := d.InspectTask(task.ID) require.NoError(err) d.(*Driver).tasks.Delete(task.ID) wg.Wait() require.True(waitDone) _, err = d.InspectTask(task.ID) require.Equal(drivers.ErrTaskNotFound, err) err = d.RecoverTask(handle) require.NoError(err) status, err := d.InspectTask(task.ID) require.NoError(err) require.Exactly(originalStatus, status) ch, err = harness.WaitTask(context.Background(), task.ID) require.NoError(err) require.NoError(d.StopTask(task.ID, 0, "SIGKILL")) select { case result := <-ch: require.NoError(result.Err) require.NotZero(result.ExitCode) // when killing a task, signal might not propagate // when executor proc.Wait() call gets "wait: no child processes" error //require.Equal(9, result.Signal) case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): require.Fail("WaitTask timeout") } require.NoError(d.DestroyTask(task.ID, false)) } // Verifies mounting a volume from the host machine and writing // some data to it from inside the container func TestRktDriver_Start_Wait_Volume(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) // enable volumes config := &Config{VolumesEnabled: true} var data []byte require.NoError(basePlug.MsgPackEncode(&data, config)) bconfig := &basePlug.Config{PluginConfig: data} require.NoError(harness.SetConfig(bconfig)) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "rkttest_alpine", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } exp := []byte{'w', 'i', 'n'} file := "output.txt" tmpvol, err := ioutil.TempDir("", "nomadtest_rktdriver_volumes") require.NoError(err) defer os.RemoveAll(tmpvol) hostpath := filepath.Join(tmpvol, file) tc := &TaskConfig{ ImageName: "docker://redis:3.2-alpine", Command: "/bin/sh", Args: []string{ "-c", fmt.Sprintf("echo -n %s > /foo/%s", string(exp), file), }, Net: []string{"none"}, Volumes: []string{fmt.Sprintf("%s:/foo", tmpvol)}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, _, err = harness.StartTask(task) require.NoError(err) // Task should terminate quickly waitCh, err := harness.WaitTask(context.Background(), task.ID) require.NoError(err) select { case res := <-waitCh: require.NoError(res.Err) require.True(res.Successful(), fmt.Sprintf("exit code %v", res.ExitCode)) case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): require.Fail("WaitTask timeout") } // Check that data was written to the shared alloc directory. act, err := ioutil.ReadFile(hostpath) require.NoError(err) require.Exactly(exp, act) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies mounting a task mount from the host machine and writing // some data to it from inside the container func TestRktDriver_Start_Wait_TaskMounts(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) // mounts through task config should be enabled regardless config := &Config{VolumesEnabled: false} var data []byte require.NoError(basePlug.MsgPackEncode(&data, config)) bconfig := &basePlug.Config{PluginConfig: data} require.NoError(harness.SetConfig(bconfig)) tmpvol, err := ioutil.TempDir("", "nomadtest_rktdriver_volumes") require.NoError(err) defer os.RemoveAll(tmpvol) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "rkttest_alpine", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, Mounts: []*drivers.MountConfig{ {HostPath: tmpvol, TaskPath: "/foo", Readonly: false}, }, } exp := []byte{'w', 'i', 'n'} file := "output.txt" hostpath := filepath.Join(tmpvol, file) tc := &TaskConfig{ ImageName: "docker://redis:3.2-alpine", Command: "/bin/sh", Args: []string{ "-c", fmt.Sprintf("echo -n %s > /foo/%s", string(exp), file), }, Net: []string{"none"}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, _, err = harness.StartTask(task) require.NoError(err) // Task should terminate quickly waitCh, err := harness.WaitTask(context.Background(), task.ID) require.NoError(err) select { case res := <-waitCh: require.NoError(res.Err) require.True(res.Successful(), fmt.Sprintf("exit code %v", res.ExitCode)) case <-time.After(time.Duration(testutil.TestMultiplier()*5) * time.Second): require.Fail("WaitTask timeout") } // Check that data was written to the shared alloc directory. act, err := ioutil.ReadFile(hostpath) require.NoError(err) require.Exactly(exp, act) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies port mapping func TestRktDriver_PortMapping(t *testing.T) { ctestutil.RktCompatible(t) require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "redis", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, Networks: []*structs.NetworkResource{ { IP: "127.0.0.1", ReservedPorts: []structs.Port{{Label: "main", Value: 8080}}, }, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ ImageName: "docker://redis:3.2-alpine", PortMap: map[string]string{ "main": "6379-tcp", }, Debug: true, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, driverNetwork, err := harness.StartTask(task) require.NoError(err) require.NotNil(driverNetwork) require.NoError(harness.DestroyTask(task.ID, true)) } // This test starts a redis container, setting user and group. // It verifies that running ps inside the container shows the expected user and group func TestRktDriver_UserGroup(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), User: "nobody", Name: "rkttest_alpine", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ ImageName: "docker://redis:3.2-alpine", Group: "nogroup", Command: "sleep", Args: []string{"9000"}, Net: []string{"none"}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, _, err := harness.StartTask(task) require.NoError(err) expected := []byte("\nnobody nogroup /bin/sleep 9000\n") testutil.WaitForResult(func() (bool, error) { res, err := d.ExecTask(task.ID, []string{"ps", "-o", "user,group,args"}, time.Second) if err != nil { return false, fmt.Errorf("failed to exec: %#v", err) } if !res.ExitResult.Successful() { return false, fmt.Errorf("ps failed: %#v %#v", res.ExitResult, res) } raw := res.Stdout return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) }, func(err error) { require.NoError(err) }) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies executing both correct and incorrect commands inside the container func TestRktDriver_Exec(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ TrustPrefix: "coreos.com/etcd", ImageName: "coreos.com/etcd:v2.0.4", Net: []string{"none"}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, _, err := harness.StartTask(task) require.NoError(err) // Run command that should succeed expected := []byte("etcd version") testutil.WaitForResult(func() (bool, error) { res, err := d.ExecTask(task.ID, []string{"/etcd", "--version"}, time.Second) if err != nil { return false, fmt.Errorf("failed to exec: %#v", err) } if !res.ExitResult.Successful() { return false, fmt.Errorf("/etcd --version failed: %#v %#v", res.ExitResult, res) } raw := res.Stdout return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) }, func(err error) { require.NoError(err) }) // Run command that should fail expected = []byte("flag provided but not defined") testutil.WaitForResult(func() (bool, error) { res, err := d.ExecTask(task.ID, []string{"/etcd", "--cgdfgdfg"}, time.Second) if err != nil { return false, fmt.Errorf("failed to exec: %#v", err) } if res.ExitResult.Successful() { return false, fmt.Errorf("/etcd --cgdfgdfg unexpected succeeded: %#v %#v", res.ExitResult, res) } raw := res.Stdout return bytes.Contains(raw, expected), fmt.Errorf("expected %q but found:\n%s", expected, raw) }, func(err error) { require.NoError(err) }) require.NoError(harness.DestroyTask(task.ID, true)) } // Verifies getting resource usage stats // TODO(preetha) figure out why stats are zero func TestRktDriver_Stats(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ TrustPrefix: "coreos.com/etcd", ImageName: "coreos.com/etcd:v2.0.4", Command: "/etcd", Net: []string{"none"}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() handle, _, err := harness.StartTask(task) require.NoError(err) // Wait for task to start _, err = harness.WaitTask(context.Background(), handle.Config.ID) require.NoError(err) // Wait until task started require.NoError(harness.WaitUntilStarted(task.ID, 1*time.Second)) ctx, cancel := context.WithCancel(context.Background()) defer cancel() statsCh, err := d.TaskStats(ctx, task.ID, time.Second*10) require.Nil(err) select { case ru := <-statsCh: //TODO(preetha) why are these zero fmt.Printf("pid map %v\n", ru.Pids) fmt.Printf("CPU:%+v Memory:%+v", ru.ResourceUsage.CpuStats, ru.ResourceUsage.MemoryStats) case <-time.After(time.Second): require.Fail("timeout receiving stats from channel") } require.NoError(harness.DestroyTask(task.ID, true)) } func TestConfig_ParseAllHCL(t *testing.T) { cfgStr := ` config { image = "docker://redis:3.2" command = "/bin/bash" args = ["-c", "echo hi"] trust_prefix = "coreos.com/etcd" dns_servers = ["8.8.8.8"] dns_search_domains = ["example.com"] net = ["network1"] port_map { http = "80-tcp" https = "443-tcp" } volumes = [ "/host-path:/container-path", ] insecure_options = ["image", "tls", "ondisk"] no_overlay = true debug = true group = "mygroup" }` expected := &TaskConfig{ ImageName: "docker://redis:3.2", Command: "/bin/bash", Args: []string{"-c", "echo hi"}, TrustPrefix: "coreos.com/etcd", DNSServers: []string{"8.8.8.8"}, DNSSearchDomains: []string{"example.com"}, Net: []string{"network1"}, PortMap: map[string]string{ "http": "80-tcp", "https": "443-tcp", }, Volumes: []string{ "/host-path:/container-path", }, InsecureOptions: []string{"image", "tls", "ondisk"}, NoOverlay: true, Debug: true, Group: "mygroup", } var tc *TaskConfig hclutils.NewConfigParser(taskConfigSpec).ParseHCL(t, cfgStr, &tc) require.EqualValues(t, expected, tc) } func TestRkt_ExecTaskStreaming(t *testing.T) { ctestutil.RktCompatible(t) if !testutil.IsCI() { t.Parallel() } require := require.New(t) d := NewRktDriver(testlog.HCLogger(t)) harness := dtestutil.NewDriverHarness(t, d) task := &drivers.TaskConfig{ ID: uuid.Generate(), AllocID: uuid.Generate(), Name: "etcd", Resources: &drivers.Resources{ NomadResources: &structs.AllocatedTaskResources{ Memory: structs.AllocatedMemoryResources{ MemoryMB: 128, }, Cpu: structs.AllocatedCpuResources{ CpuShares: 100, }, }, LinuxResources: &drivers.LinuxResources{ MemoryLimitBytes: 134217728, CPUShares: 100, }, }, } tc := &TaskConfig{ ImageName: "docker://busybox:1.29.3", Command: "/bin/sleep", Args: []string{"1000"}, Net: []string{"none"}, } require.NoError(task.EncodeConcreteDriverConfig(&tc)) testtask.SetTaskConfigEnv(task) cleanup := harness.MkAllocDir(task, true) defer cleanup() _, _, err := harness.StartTask(task) require.NoError(err) defer d.DestroyTask(task.ID, true) // wait for container to be up and executable testutil.WaitForResult(func() (bool, error) { res, err := d.ExecTask(task.ID, []string{"/bin/sh", "-c", "echo hi"}, time.Second) if err != nil { return false, fmt.Errorf("failed to exec: %#v", err) } if !res.ExitResult.Successful() { return false, fmt.Errorf("ps failed: %#v %#v", res.ExitResult, res) } return true, nil }, func(err error) { require.NoError(err) }) dtestutil.ExecTaskStreamingConformanceTests(t, harness, task.ID) }