Merge branch 'master' into d-vault
This commit is contained in:
commit
83106de217
|
@ -536,7 +536,7 @@ func TestTaskRunner_RestartTask(t *testing.T) {
|
||||||
task.Driver = "mock_driver"
|
task.Driver = "mock_driver"
|
||||||
task.Config = map[string]interface{}{
|
task.Config = map[string]interface{}{
|
||||||
"exit_code": "0",
|
"exit_code": "0",
|
||||||
"run_for": "10s",
|
"run_for": "100s",
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx := testTaskRunnerFromAlloc(t, true, alloc)
|
ctx := testTaskRunnerFromAlloc(t, true, alloc)
|
||||||
|
|
|
@ -51,7 +51,7 @@ Dispatch Options:
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *JobDispatchCommand) Synopsis() string {
|
func (c *JobDispatchCommand) Synopsis() string {
|
||||||
return "Dispatch an instance of a parametereized job"
|
return "Dispatch an instance of a parameterized job"
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *JobDispatchCommand) Run(args []string) int {
|
func (c *JobDispatchCommand) Run(args []string) int {
|
||||||
|
|
|
@ -192,6 +192,8 @@ func (c *CoreScheduler) evalGC(eval *structs.Evaluation) error {
|
||||||
|
|
||||||
// The Evaluation GC should not handle batch jobs since those need to be
|
// The Evaluation GC should not handle batch jobs since those need to be
|
||||||
// garbage collected in one shot
|
// garbage collected in one shot
|
||||||
|
// XXX believe there is a bug that if a batch job gets stopped, there is no
|
||||||
|
// way for it to GC the eval/allocs
|
||||||
gc, allocs, err := c.gcEval(eval, oldThreshold, false)
|
gc, allocs, err := c.gcEval(eval, oldThreshold, false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
|
|
@ -965,6 +965,66 @@ func TestCoreScheduler_JobGC_Force(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// This test ensures parameterized and periodic jobs don't get GCd
|
||||||
|
func TestCoreScheduler_JobGC_NonGCable(t *testing.T) {
|
||||||
|
s1 := testServer(t, nil)
|
||||||
|
defer s1.Shutdown()
|
||||||
|
testutil.WaitForLeader(t, s1.RPC)
|
||||||
|
|
||||||
|
// COMPAT Remove in 0.6: Reset the FSM time table since we reconcile which sets index 0
|
||||||
|
s1.fsm.timetable.table = make([]TimeTableEntry, 1, 10)
|
||||||
|
|
||||||
|
// Insert a parameterized job.
|
||||||
|
state := s1.fsm.State()
|
||||||
|
job := mock.Job()
|
||||||
|
job.Type = structs.JobTypeBatch
|
||||||
|
job.Status = structs.JobStatusRunning
|
||||||
|
job.ParameterizedJob = &structs.ParameterizedJobConfig{
|
||||||
|
Payload: structs.DispatchPayloadRequired,
|
||||||
|
}
|
||||||
|
err := state.UpsertJob(1000, job)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Insert a periodic job.
|
||||||
|
job2 := mock.PeriodicJob()
|
||||||
|
if err := state.UpsertJob(1001, job2); err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a core scheduler
|
||||||
|
snap, err := state.Snapshot()
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
core := NewCoreScheduler(s1, snap)
|
||||||
|
|
||||||
|
// Attempt the GC
|
||||||
|
gc := s1.coreJobEval(structs.CoreJobForceGC, 1002)
|
||||||
|
err = core.Process(gc)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Should still exist
|
||||||
|
out, err := state.JobByID(job.ID)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
if out == nil {
|
||||||
|
t.Fatalf("bad: %v", out)
|
||||||
|
}
|
||||||
|
|
||||||
|
outE, err := state.JobByID(job2.ID)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("err: %v", err)
|
||||||
|
}
|
||||||
|
if outE == nil {
|
||||||
|
t.Fatalf("bad: %v", outE)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func TestCoreScheduler_PartitionReap(t *testing.T) {
|
func TestCoreScheduler_PartitionReap(t *testing.T) {
|
||||||
s1 := testServer(t, nil)
|
s1 := testServer(t, nil)
|
||||||
defer s1.Shutdown()
|
defer s1.Shutdown()
|
||||||
|
|
|
@ -70,7 +70,7 @@ func TestResetHeartbeatTimerLocked(t *testing.T) {
|
||||||
t.Fatalf("missing timer")
|
t.Fatalf("missing timer")
|
||||||
}
|
}
|
||||||
|
|
||||||
time.Sleep(10 * time.Millisecond)
|
time.Sleep(time.Duration(testutil.TestMultiplier()*10) * time.Millisecond)
|
||||||
|
|
||||||
if _, ok := s1.heartbeatTimers["foo"]; ok {
|
if _, ok := s1.heartbeatTimers["foo"]; ok {
|
||||||
t.Fatalf("timer should be gone")
|
t.Fatalf("timer should be gone")
|
||||||
|
@ -99,7 +99,7 @@ func TestResetHeartbeatTimerLocked_Renew(t *testing.T) {
|
||||||
renew := time.Now()
|
renew := time.Now()
|
||||||
|
|
||||||
// Watch for invalidation
|
// Watch for invalidation
|
||||||
for time.Now().Sub(renew) < 20*time.Millisecond {
|
for time.Now().Sub(renew) < time.Duration(testutil.TestMultiplier()*20)*time.Millisecond {
|
||||||
s1.heartbeatTimersLock.Lock()
|
s1.heartbeatTimersLock.Lock()
|
||||||
_, ok := s1.heartbeatTimers["foo"]
|
_, ok := s1.heartbeatTimers["foo"]
|
||||||
s1.heartbeatTimersLock.Unlock()
|
s1.heartbeatTimersLock.Unlock()
|
||||||
|
|
|
@ -149,9 +149,10 @@ func jobIsGCable(obj interface{}) (bool, error) {
|
||||||
return false, fmt.Errorf("Unexpected type: %v", obj)
|
return false, fmt.Errorf("Unexpected type: %v", obj)
|
||||||
}
|
}
|
||||||
|
|
||||||
// The job is GCable if it is batch and it is not periodic
|
// The job is GCable if it is batch, it is not periodic and is not a
|
||||||
|
// parameterized job.
|
||||||
periodic := j.Periodic != nil && j.Periodic.Enabled
|
periodic := j.Periodic != nil && j.Periodic.Enabled
|
||||||
gcable := j.Type == structs.JobTypeBatch && !periodic
|
gcable := j.Type == structs.JobTypeBatch && !periodic && !j.IsParameterized()
|
||||||
return gcable, nil
|
return gcable, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,6 @@
|
||||||
package config
|
package config
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/tls"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"net/http"
|
"net/http"
|
||||||
"strings"
|
"strings"
|
||||||
|
@ -121,7 +120,7 @@ func (a *ConsulConfig) Merge(b *ConsulConfig) *ConsulConfig {
|
||||||
result.EnableSSL = b.EnableSSL
|
result.EnableSSL = b.EnableSSL
|
||||||
}
|
}
|
||||||
if b.VerifySSL != nil {
|
if b.VerifySSL != nil {
|
||||||
result.VerifySSL = b.EnableSSL
|
result.VerifySSL = b.VerifySSL
|
||||||
}
|
}
|
||||||
if b.CAFile != "" {
|
if b.CAFile != "" {
|
||||||
result.CAFile = b.CAFile
|
result.CAFile = b.CAFile
|
||||||
|
@ -180,6 +179,7 @@ func (c *ConsulConfig) ApiConfig() (*consul.Config, error) {
|
||||||
if c.VerifySSL != nil {
|
if c.VerifySSL != nil {
|
||||||
tlsConfig.InsecureSkipVerify = !*c.VerifySSL
|
tlsConfig.InsecureSkipVerify = !*c.VerifySSL
|
||||||
}
|
}
|
||||||
|
|
||||||
tlsClientCfg, err := consul.SetupTLSConfig(&tlsConfig)
|
tlsClientCfg, err := consul.SetupTLSConfig(&tlsConfig)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("error creating tls client config for consul: %v", err)
|
return nil, fmt.Errorf("error creating tls client config for consul: %v", err)
|
||||||
|
@ -188,13 +188,6 @@ func (c *ConsulConfig) ApiConfig() (*consul.Config, error) {
|
||||||
TLSClientConfig: tlsClientCfg,
|
TLSClientConfig: tlsClientCfg,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if c.EnableSSL != nil && !*c.VerifySSL {
|
|
||||||
config.HttpClient.Transport = &http.Transport{
|
|
||||||
TLSClientConfig: &tls.Config{
|
|
||||||
InsecureSkipVerify: true,
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return config, nil
|
return config, nil
|
||||||
}
|
}
|
||||||
|
|
|
@ -591,9 +591,6 @@ func TestVaultClient_LookupToken_RateLimit(t *testing.T) {
|
||||||
unblock := make(chan struct{})
|
unblock := make(chan struct{})
|
||||||
for i := 0; i < numRequests; i++ {
|
for i := 0; i < numRequests; i++ {
|
||||||
go func() {
|
go func() {
|
||||||
// Ensure all the goroutines are made
|
|
||||||
time.Sleep(10 * time.Millisecond)
|
|
||||||
|
|
||||||
// Lookup ourselves
|
// Lookup ourselves
|
||||||
_, err := client.LookupToken(ctx, v.Config.Token)
|
_, err := client.LookupToken(ctx, v.Config.Token)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -607,7 +604,7 @@ func TestVaultClient_LookupToken_RateLimit(t *testing.T) {
|
||||||
|
|
||||||
// Cancel the context
|
// Cancel the context
|
||||||
cancel()
|
cancel()
|
||||||
time.AfterFunc(1*time.Second, func() { close(unblock) })
|
close(unblock)
|
||||||
}()
|
}()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -618,9 +615,15 @@ func TestVaultClient_LookupToken_RateLimit(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
desired := numRequests - 1
|
desired := numRequests - 1
|
||||||
|
testutil.WaitForResult(func() (bool, error) {
|
||||||
if cancels != desired {
|
if cancels != desired {
|
||||||
t.Fatalf("Incorrect number of cancels; got %d; want %d", cancels, desired)
|
return false, fmt.Errorf("Incorrect number of cancels; got %d; want %d", cancels, desired)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
return true, nil
|
||||||
|
}, func(err error) {
|
||||||
|
t.Fatalf("Connection not established")
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestVaultClient_CreateToken_Root(t *testing.T) {
|
func TestVaultClient_CreateToken_Root(t *testing.T) {
|
||||||
|
|
|
@ -54,3 +54,18 @@ are not participating in Raft. Thus clients can have 100+ millisecond latency to
|
||||||
their servers. This allows having a set of Nomad servers that service clients
|
their servers. This allows having a set of Nomad servers that service clients
|
||||||
that can be spread geographically over a continent or even the world in the case
|
that can be spread geographically over a continent or even the world in the case
|
||||||
of having a single "global" region and many datacenter.
|
of having a single "global" region and many datacenter.
|
||||||
|
|
||||||
|
## Ports Used
|
||||||
|
|
||||||
|
Nomad requires 3 different ports to work properly on servers and 2 on clients,
|
||||||
|
some on TCP, UDP, or both protocols. Below we document the requirements for each
|
||||||
|
port.
|
||||||
|
|
||||||
|
* HTTP API (Default 4646). This is used by clients and servers to serve the HTTP
|
||||||
|
API. TCP only.
|
||||||
|
|
||||||
|
* RPC (Default 4647). This is used by servers and clients to communicate amongst
|
||||||
|
each other. TCP only.
|
||||||
|
|
||||||
|
* Serf WAN (Default 4648). This is used by servers to gossip over the WAN to
|
||||||
|
other servers. TCP and UDP.
|
||||||
|
|
Loading…
Reference in a new issue