Merge pull request #10806 from hashicorp/dnephin/debug-filenames-2
debug: use human readable dates for filenames and improve the tests
This commit is contained in:
commit
17c9bac789
|
@ -18,10 +18,9 @@ import (
|
||||||
"syscall"
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"golang.org/x/sync/errgroup"
|
|
||||||
|
|
||||||
"github.com/hashicorp/go-multierror"
|
"github.com/hashicorp/go-multierror"
|
||||||
"github.com/mitchellh/cli"
|
"github.com/mitchellh/cli"
|
||||||
|
"golang.org/x/sync/errgroup"
|
||||||
|
|
||||||
"github.com/hashicorp/consul/api"
|
"github.com/hashicorp/consul/api"
|
||||||
"github.com/hashicorp/consul/command/flags"
|
"github.com/hashicorp/consul/command/flags"
|
||||||
|
@ -86,8 +85,9 @@ type cmd struct {
|
||||||
// validateTiming can be used to skip validation of interval, duration. This
|
// validateTiming can be used to skip validation of interval, duration. This
|
||||||
// is primarily useful for testing
|
// is primarily useful for testing
|
||||||
validateTiming bool
|
validateTiming bool
|
||||||
|
// timeNow is a shim for testing, it is used to generate the time used in
|
||||||
index *debugIndex
|
// file paths.
|
||||||
|
timeNow func() time.Time
|
||||||
}
|
}
|
||||||
|
|
||||||
// debugIndex is used to manage the summary of all data recorded
|
// debugIndex is used to manage the summary of all data recorded
|
||||||
|
@ -105,10 +105,16 @@ type debugIndex struct {
|
||||||
Targets []string
|
Targets []string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// timeDateformat is a modified version of time.RFC3339 which replaces colons with
|
||||||
|
// hyphens. This is to make it more convenient to untar these files, because
|
||||||
|
// tar assumes colons indicate the file is on a remote host, unless --force-local
|
||||||
|
// is used.
|
||||||
|
const timeDateFormat = "2006-01-02T15-04-05Z0700"
|
||||||
|
|
||||||
func (c *cmd) init() {
|
func (c *cmd) init() {
|
||||||
c.flags = flag.NewFlagSet("", flag.ContinueOnError)
|
c.flags = flag.NewFlagSet("", flag.ContinueOnError)
|
||||||
|
|
||||||
defaultFilename := fmt.Sprintf("consul-debug-%d", time.Now().Unix())
|
defaultFilename := fmt.Sprintf("consul-debug-%v", time.Now().Format(timeDateFormat))
|
||||||
|
|
||||||
c.flags.Var((*flags.AppendSliceValue)(&c.capture), "capture",
|
c.flags.Var((*flags.AppendSliceValue)(&c.capture), "capture",
|
||||||
fmt.Sprintf("One or more types of information to capture. This can be used "+
|
fmt.Sprintf("One or more types of information to capture. This can be used "+
|
||||||
|
@ -133,6 +139,9 @@ func (c *cmd) init() {
|
||||||
c.help = flags.Usage(help, c.flags)
|
c.help = flags.Usage(help, c.flags)
|
||||||
|
|
||||||
c.validateTiming = true
|
c.validateTiming = true
|
||||||
|
c.timeNow = func() time.Time {
|
||||||
|
return time.Now().UTC()
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) Run(args []string) int {
|
func (c *cmd) Run(args []string) int {
|
||||||
|
@ -178,15 +187,6 @@ func (c *cmd) Run(args []string) int {
|
||||||
c.UI.Info(fmt.Sprintf(" Output: '%s'", archiveName))
|
c.UI.Info(fmt.Sprintf(" Output: '%s'", archiveName))
|
||||||
c.UI.Info(fmt.Sprintf(" Capture: '%s'", strings.Join(c.capture, ", ")))
|
c.UI.Info(fmt.Sprintf(" Capture: '%s'", strings.Join(c.capture, ", ")))
|
||||||
|
|
||||||
// Record some information for the index at the root of the archive
|
|
||||||
index := &debugIndex{
|
|
||||||
Version: debugProtocolVersion,
|
|
||||||
AgentVersion: version,
|
|
||||||
Interval: c.interval.String(),
|
|
||||||
Duration: c.duration.String(),
|
|
||||||
Targets: c.capture,
|
|
||||||
}
|
|
||||||
|
|
||||||
// Add the extra grace period to ensure
|
// Add the extra grace period to ensure
|
||||||
// all intervals will be captured within the time allotted
|
// all intervals will be captured within the time allotted
|
||||||
c.duration = c.duration + debugDurationGrace
|
c.duration = c.duration + debugDurationGrace
|
||||||
|
@ -212,15 +212,15 @@ func (c *cmd) Run(args []string) int {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Write the index document
|
// Record some information for the index at the root of the archive
|
||||||
idxMarshalled, err := json.MarshalIndent(index, "", "\t")
|
index := &debugIndex{
|
||||||
if err != nil {
|
Version: debugProtocolVersion,
|
||||||
c.UI.Error(fmt.Sprintf("Error marshalling index document: %v", err))
|
AgentVersion: version,
|
||||||
return 1
|
Interval: c.interval.String(),
|
||||||
|
Duration: c.duration.String(),
|
||||||
|
Targets: c.capture,
|
||||||
}
|
}
|
||||||
|
if err := writeJSONFile(filepath.Join(c.output, "index.json"), index); err != nil {
|
||||||
err = ioutil.WriteFile(fmt.Sprintf("%s/index.json", c.output), idxMarshalled, 0644)
|
|
||||||
if err != nil {
|
|
||||||
c.UI.Error(fmt.Sprintf("Error creating index document: %v", err))
|
c.UI.Error(fmt.Sprintf("Error creating index document: %v", err))
|
||||||
return 1
|
return 1
|
||||||
}
|
}
|
||||||
|
@ -346,13 +346,13 @@ func (c *cmd) captureInterval(ctx context.Context) error {
|
||||||
durationChn := time.After(c.duration)
|
durationChn := time.After(c.duration)
|
||||||
intervalCount := 0
|
intervalCount := 0
|
||||||
|
|
||||||
c.UI.Output(fmt.Sprintf("Beginning capture interval %s (%d)", time.Now().Local().String(), intervalCount))
|
c.UI.Output(fmt.Sprintf("Beginning capture interval %s (%d)", time.Now().UTC(), intervalCount))
|
||||||
|
|
||||||
err := captureShortLived(c)
|
err := captureShortLived(c)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
c.UI.Output(fmt.Sprintf("Capture successful %s (%d)", time.Now().Local().String(), intervalCount))
|
c.UI.Output(fmt.Sprintf("Capture successful %s (%d)", time.Now().UTC(), intervalCount))
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case t := <-intervalChn.C:
|
case t := <-intervalChn.C:
|
||||||
|
@ -361,7 +361,7 @@ func (c *cmd) captureInterval(ctx context.Context) error {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
c.UI.Output(fmt.Sprintf("Capture successful %s (%d)", t.Local().String(), intervalCount))
|
c.UI.Output(fmt.Sprintf("Capture successful %s (%d)", t.UTC(), intervalCount))
|
||||||
case <-durationChn:
|
case <-durationChn:
|
||||||
intervalChn.Stop()
|
intervalChn.Stop()
|
||||||
return nil
|
return nil
|
||||||
|
@ -373,43 +373,32 @@ func (c *cmd) captureInterval(ctx context.Context) error {
|
||||||
|
|
||||||
func captureShortLived(c *cmd) error {
|
func captureShortLived(c *cmd) error {
|
||||||
g := new(errgroup.Group)
|
g := new(errgroup.Group)
|
||||||
timestamp := time.Now().Local().Unix()
|
|
||||||
|
|
||||||
timestampDir, err := c.createTimestampDir(timestamp)
|
dir, err := makeIntervalDir(c.output, c.timeNow())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if c.captureTarget(targetProfiles) {
|
if c.captureTarget(targetProfiles) {
|
||||||
g.Go(func() error {
|
g.Go(func() error {
|
||||||
return c.captureHeap(timestampDir)
|
return c.captureHeap(dir)
|
||||||
})
|
})
|
||||||
|
|
||||||
g.Go(func() error {
|
g.Go(func() error {
|
||||||
return c.captureGoRoutines(timestampDir)
|
return c.captureGoRoutines(dir)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
return g.Wait()
|
return g.Wait()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) createTimestampDir(timestamp int64) (string, error) {
|
func makeIntervalDir(base string, now time.Time) (string, error) {
|
||||||
// Make the directory that will store all captured data
|
dir := filepath.Join(base, now.Format(timeDateFormat))
|
||||||
// for this interval
|
if err := os.MkdirAll(dir, 0755); err != nil {
|
||||||
timestampDir := fmt.Sprintf("%s/%d", c.output, timestamp)
|
return "", fmt.Errorf("failed to create output directory %v: %w", dir, err)
|
||||||
err := os.MkdirAll(timestampDir, 0755)
|
|
||||||
if err != nil {
|
|
||||||
return "", err
|
|
||||||
}
|
}
|
||||||
return timestampDir, nil
|
return dir, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureLongRunning(ctx context.Context) error {
|
func (c *cmd) captureLongRunning(ctx context.Context) error {
|
||||||
timestamp := time.Now().Local().Unix()
|
|
||||||
|
|
||||||
timestampDir, err := c.createTimestampDir(timestamp)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
g := new(errgroup.Group)
|
g := new(errgroup.Group)
|
||||||
// Capture a profile/trace with a minimum of 1s
|
// Capture a profile/trace with a minimum of 1s
|
||||||
s := c.duration.Seconds()
|
s := c.duration.Seconds()
|
||||||
|
@ -419,43 +408,42 @@ func (c *cmd) captureLongRunning(ctx context.Context) error {
|
||||||
if c.captureTarget(targetProfiles) {
|
if c.captureTarget(targetProfiles) {
|
||||||
g.Go(func() error {
|
g.Go(func() error {
|
||||||
// use ctx without a timeout to allow the profile to finish sending
|
// use ctx without a timeout to allow the profile to finish sending
|
||||||
return c.captureProfile(ctx, s, timestampDir)
|
return c.captureProfile(ctx, s)
|
||||||
})
|
})
|
||||||
|
|
||||||
g.Go(func() error {
|
g.Go(func() error {
|
||||||
// use ctx without a timeout to allow the trace to finish sending
|
// use ctx without a timeout to allow the trace to finish sending
|
||||||
return c.captureTrace(ctx, s, timestampDir)
|
return c.captureTrace(ctx, s)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
if c.captureTarget(targetLogs) {
|
if c.captureTarget(targetLogs) {
|
||||||
g.Go(func() error {
|
g.Go(func() error {
|
||||||
ctx, cancel := context.WithTimeout(ctx, c.duration)
|
ctx, cancel := context.WithTimeout(ctx, c.duration)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
return c.captureLogs(ctx, timestampDir)
|
return c.captureLogs(ctx)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
if c.captureTarget(targetMetrics) {
|
if c.captureTarget(targetMetrics) {
|
||||||
g.Go(func() error {
|
g.Go(func() error {
|
||||||
|
|
||||||
ctx, cancel := context.WithTimeout(ctx, c.duration)
|
ctx, cancel := context.WithTimeout(ctx, c.duration)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
return c.captureMetrics(ctx, timestampDir)
|
return c.captureMetrics(ctx)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
return g.Wait()
|
return g.Wait()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureGoRoutines(timestampDir string) error {
|
func (c *cmd) captureGoRoutines(outputDir string) error {
|
||||||
gr, err := c.client.Debug().Goroutine()
|
gr, err := c.client.Debug().Goroutine()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to collect goroutine profile: %w", err)
|
return fmt.Errorf("failed to collect goroutine profile: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return ioutil.WriteFile(fmt.Sprintf("%s/goroutine.prof", timestampDir), gr, 0644)
|
return ioutil.WriteFile(filepath.Join(outputDir, "goroutine.prof"), gr, 0644)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureTrace(ctx context.Context, s float64, timestampDir string) error {
|
func (c *cmd) captureTrace(ctx context.Context, s float64) error {
|
||||||
prof, err := c.client.Debug().PProf(ctx, "trace", int(s))
|
prof, err := c.client.Debug().PProf(ctx, "trace", int(s))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to collect cpu profile: %w", err)
|
return fmt.Errorf("failed to collect cpu profile: %w", err)
|
||||||
|
@ -463,7 +451,7 @@ func (c *cmd) captureTrace(ctx context.Context, s float64, timestampDir string)
|
||||||
defer prof.Close()
|
defer prof.Close()
|
||||||
|
|
||||||
r := bufio.NewReader(prof)
|
r := bufio.NewReader(prof)
|
||||||
fh, err := os.Create(fmt.Sprintf("%s/trace.out", timestampDir))
|
fh, err := os.Create(filepath.Join(c.output, "trace.out"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -472,7 +460,7 @@ func (c *cmd) captureTrace(ctx context.Context, s float64, timestampDir string)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureProfile(ctx context.Context, s float64, timestampDir string) error {
|
func (c *cmd) captureProfile(ctx context.Context, s float64) error {
|
||||||
prof, err := c.client.Debug().PProf(ctx, "profile", int(s))
|
prof, err := c.client.Debug().PProf(ctx, "profile", int(s))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to collect cpu profile: %w", err)
|
return fmt.Errorf("failed to collect cpu profile: %w", err)
|
||||||
|
@ -480,7 +468,7 @@ func (c *cmd) captureProfile(ctx context.Context, s float64, timestampDir string
|
||||||
defer prof.Close()
|
defer prof.Close()
|
||||||
|
|
||||||
r := bufio.NewReader(prof)
|
r := bufio.NewReader(prof)
|
||||||
fh, err := os.Create(fmt.Sprintf("%s/profile.prof", timestampDir))
|
fh, err := os.Create(filepath.Join(c.output, "profile.prof"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -489,23 +477,23 @@ func (c *cmd) captureProfile(ctx context.Context, s float64, timestampDir string
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureHeap(timestampDir string) error {
|
func (c *cmd) captureHeap(outputDir string) error {
|
||||||
heap, err := c.client.Debug().Heap()
|
heap, err := c.client.Debug().Heap()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to collect heap profile: %w", err)
|
return fmt.Errorf("failed to collect heap profile: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return ioutil.WriteFile(fmt.Sprintf("%s/heap.prof", timestampDir), heap, 0644)
|
return ioutil.WriteFile(filepath.Join(outputDir, "heap.prof"), heap, 0644)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureLogs(ctx context.Context, timestampDir string) error {
|
func (c *cmd) captureLogs(ctx context.Context) error {
|
||||||
logCh, err := c.client.Agent().Monitor("DEBUG", ctx.Done(), nil)
|
logCh, err := c.client.Agent().Monitor("DEBUG", ctx.Done(), nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create the log file for writing
|
// Create the log file for writing
|
||||||
f, err := os.Create(fmt.Sprintf("%s/%s", timestampDir, "consul.log"))
|
f, err := os.Create(filepath.Join(c.output, "consul.log"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -526,15 +514,14 @@ func (c *cmd) captureLogs(ctx context.Context, timestampDir string) error {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *cmd) captureMetrics(ctx context.Context, timestampDir string) error {
|
func (c *cmd) captureMetrics(ctx context.Context) error {
|
||||||
stream, err := c.client.Agent().MetricsStream(ctx)
|
stream, err := c.client.Agent().MetricsStream(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer stream.Close()
|
defer stream.Close()
|
||||||
|
|
||||||
filename := fmt.Sprintf("%s/%s.json", timestampDir, "metrics")
|
fh, err := os.Create(filepath.Join(c.output, "metrics.json"))
|
||||||
fh, err := os.Create(filename)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to create metrics file: %w", err)
|
return fmt.Errorf("failed to create metrics file: %w", err)
|
||||||
}
|
}
|
||||||
|
|
|
@ -3,10 +3,11 @@ package debug
|
||||||
import (
|
import (
|
||||||
"archive/tar"
|
"archive/tar"
|
||||||
"bufio"
|
"bufio"
|
||||||
|
"bytes"
|
||||||
"compress/gzip"
|
"compress/gzip"
|
||||||
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"regexp"
|
"regexp"
|
||||||
|
@ -18,6 +19,7 @@ import (
|
||||||
"github.com/mitchellh/cli"
|
"github.com/mitchellh/cli"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
"gotest.tools/v3/assert"
|
"gotest.tools/v3/assert"
|
||||||
|
"gotest.tools/v3/assert/cmp"
|
||||||
"gotest.tools/v3/fs"
|
"gotest.tools/v3/fs"
|
||||||
|
|
||||||
"github.com/hashicorp/consul/agent"
|
"github.com/hashicorp/consul/agent"
|
||||||
|
@ -49,12 +51,17 @@ func TestDebugCommand(t *testing.T) {
|
||||||
cmd := New(ui)
|
cmd := New(ui)
|
||||||
cmd.validateTiming = false
|
cmd.validateTiming = false
|
||||||
|
|
||||||
|
it := &incrementalTime{
|
||||||
|
base: time.Date(2021, 7, 8, 9, 10, 11, 0, time.UTC),
|
||||||
|
}
|
||||||
|
cmd.timeNow = it.Now
|
||||||
|
|
||||||
outputPath := fmt.Sprintf("%s/debug", testDir)
|
outputPath := fmt.Sprintf("%s/debug", testDir)
|
||||||
args := []string{
|
args := []string{
|
||||||
"-http-addr=" + a.HTTPAddr(),
|
"-http-addr=" + a.HTTPAddr(),
|
||||||
"-output=" + outputPath,
|
"-output=" + outputPath,
|
||||||
"-duration=100ms",
|
"-duration=2s",
|
||||||
"-interval=50ms",
|
"-interval=1s",
|
||||||
"-archive=false",
|
"-archive=false",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -64,17 +71,75 @@ func TestDebugCommand(t *testing.T) {
|
||||||
|
|
||||||
expected := fs.Expected(t,
|
expected := fs.Expected(t,
|
||||||
fs.WithDir("debug",
|
fs.WithDir("debug",
|
||||||
fs.WithFile("agent.json", "", fs.MatchAnyFileContent),
|
fs.WithFile("index.json", "", fs.MatchFileContent(validIndexJSON)),
|
||||||
fs.WithFile("host.json", "", fs.MatchAnyFileContent),
|
fs.WithFile("agent.json", "", fs.MatchFileContent(validJSON)),
|
||||||
fs.WithFile("index.json", "", fs.MatchAnyFileContent),
|
fs.WithFile("host.json", "", fs.MatchFileContent(validJSON)),
|
||||||
fs.WithFile("members.json", "", fs.MatchAnyFileContent),
|
fs.WithFile("members.json", "", fs.MatchFileContent(validJSON)),
|
||||||
// TODO: make the sub-directory names predictable)
|
fs.WithFile("metrics.json", "", fs.MatchAnyFileContent),
|
||||||
|
fs.WithFile("consul.log", "", fs.MatchFileContent(validLogFile)),
|
||||||
|
fs.WithFile("profile.prof", "", fs.MatchFileContent(validProfileData)),
|
||||||
|
fs.WithFile("trace.out", "", fs.MatchAnyFileContent),
|
||||||
|
fs.WithDir("2021-07-08T09-10-12Z",
|
||||||
|
fs.WithFile("goroutine.prof", "", fs.MatchFileContent(validProfileData)),
|
||||||
|
fs.WithFile("heap.prof", "", fs.MatchFileContent(validProfileData))),
|
||||||
|
fs.WithDir("2021-07-08T09-10-13Z",
|
||||||
|
fs.WithFile("goroutine.prof", "", fs.MatchFileContent(validProfileData)),
|
||||||
|
fs.WithFile("heap.prof", "", fs.MatchFileContent(validProfileData))),
|
||||||
|
// Ignore the extra directories, they should be the same as the first two
|
||||||
fs.MatchExtraFiles))
|
fs.MatchExtraFiles))
|
||||||
assert.Assert(t, fs.Equal(testDir, expected))
|
assert.Assert(t, fs.Equal(testDir, expected))
|
||||||
|
|
||||||
metricsFiles, err := filepath.Glob(fmt.Sprintf("%s/*/%s", outputPath, "metrics.json"))
|
require.Equal(t, "", ui.ErrorWriter.String(), "expected no error output")
|
||||||
require.NoError(t, err)
|
}
|
||||||
require.Len(t, metricsFiles, 1)
|
|
||||||
|
func validLogFile(raw []byte) fs.CompareResult {
|
||||||
|
scanner := bufio.NewScanner(bytes.NewReader(raw))
|
||||||
|
for scanner.Scan() {
|
||||||
|
logLine := scanner.Text()
|
||||||
|
if !validateLogLine([]byte(logLine)) {
|
||||||
|
return cmp.ResultFailure(fmt.Sprintf("log line is not valid %s", logLine))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if scanner.Err() != nil {
|
||||||
|
return cmp.ResultFailure(scanner.Err().Error())
|
||||||
|
}
|
||||||
|
return cmp.ResultSuccess
|
||||||
|
}
|
||||||
|
|
||||||
|
func validIndexJSON(raw []byte) fs.CompareResult {
|
||||||
|
var target debugIndex
|
||||||
|
decoder := json.NewDecoder(bytes.NewReader(raw))
|
||||||
|
decoder.DisallowUnknownFields()
|
||||||
|
if err := decoder.Decode(&target); err != nil {
|
||||||
|
return cmp.ResultFailure(err.Error())
|
||||||
|
}
|
||||||
|
return cmp.ResultSuccess
|
||||||
|
}
|
||||||
|
|
||||||
|
func validJSON(raw []byte) fs.CompareResult {
|
||||||
|
var target interface{}
|
||||||
|
decoder := json.NewDecoder(bytes.NewReader(raw))
|
||||||
|
if err := decoder.Decode(&target); err != nil {
|
||||||
|
return cmp.ResultFailure(err.Error())
|
||||||
|
}
|
||||||
|
return cmp.ResultSuccess
|
||||||
|
}
|
||||||
|
|
||||||
|
func validProfileData(raw []byte) fs.CompareResult {
|
||||||
|
if _, err := profile.ParseData(raw); err != nil {
|
||||||
|
return cmp.ResultFailure(err.Error())
|
||||||
|
}
|
||||||
|
return cmp.ResultSuccess
|
||||||
|
}
|
||||||
|
|
||||||
|
type incrementalTime struct {
|
||||||
|
base time.Time
|
||||||
|
next uint64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *incrementalTime) Now() time.Time {
|
||||||
|
t.next++
|
||||||
|
return t.base.Add(time.Duration(t.next) * time.Second)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDebugCommand_Archive(t *testing.T) {
|
func TestDebugCommand_Archive(t *testing.T) {
|
||||||
|
@ -267,11 +332,11 @@ func TestDebugCommand_CaptureTargets(t *testing.T) {
|
||||||
"static": {
|
"static": {
|
||||||
[]string{"agent", "host", "cluster"},
|
[]string{"agent", "host", "cluster"},
|
||||||
[]string{"agent.json", "host.json", "members.json"},
|
[]string{"agent.json", "host.json", "members.json"},
|
||||||
[]string{"*/metrics.json"},
|
[]string{"metrics.json"},
|
||||||
},
|
},
|
||||||
"metrics-only": {
|
"metrics-only": {
|
||||||
[]string{"metrics"},
|
[]string{"metrics"},
|
||||||
[]string{"*/metrics.json"},
|
[]string{"metrics.json"},
|
||||||
[]string{"agent.json", "host.json", "members.json"},
|
[]string{"agent.json", "host.json", "members.json"},
|
||||||
},
|
},
|
||||||
"all-but-pprof": {
|
"all-but-pprof": {
|
||||||
|
@ -286,8 +351,8 @@ func TestDebugCommand_CaptureTargets(t *testing.T) {
|
||||||
"host.json",
|
"host.json",
|
||||||
"agent.json",
|
"agent.json",
|
||||||
"members.json",
|
"members.json",
|
||||||
"*/metrics.json",
|
"metrics.json",
|
||||||
"*/consul.log",
|
"consul.log",
|
||||||
},
|
},
|
||||||
[]string{},
|
[]string{},
|
||||||
},
|
},
|
||||||
|
@ -356,100 +421,6 @@ func TestDebugCommand_CaptureTargets(t *testing.T) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDebugCommand_CaptureLogs(t *testing.T) {
|
|
||||||
if testing.Short() {
|
|
||||||
t.Skip("too slow for testing.Short")
|
|
||||||
}
|
|
||||||
|
|
||||||
cases := map[string]struct {
|
|
||||||
// used in -target param
|
|
||||||
targets []string
|
|
||||||
// existence verified after execution
|
|
||||||
files []string
|
|
||||||
// non-existence verified after execution
|
|
||||||
excludedFiles []string
|
|
||||||
}{
|
|
||||||
"logs-only": {
|
|
||||||
[]string{"logs"},
|
|
||||||
[]string{"*/consul.log"},
|
|
||||||
[]string{"agent.json", "host.json", "cluster.json", "*/metrics.json"},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for name, tc := range cases {
|
|
||||||
testDir := testutil.TempDir(t, "debug")
|
|
||||||
|
|
||||||
a := agent.NewTestAgent(t, `
|
|
||||||
enable_debug = true
|
|
||||||
`)
|
|
||||||
|
|
||||||
defer a.Shutdown()
|
|
||||||
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
|
||||||
|
|
||||||
ui := cli.NewMockUi()
|
|
||||||
cmd := New(ui)
|
|
||||||
cmd.validateTiming = false
|
|
||||||
|
|
||||||
outputPath := fmt.Sprintf("%s/debug-%s", testDir, name)
|
|
||||||
args := []string{
|
|
||||||
"-http-addr=" + a.HTTPAddr(),
|
|
||||||
"-output=" + outputPath,
|
|
||||||
"-archive=false",
|
|
||||||
"-duration=1000ms",
|
|
||||||
"-interval=50ms",
|
|
||||||
}
|
|
||||||
for _, t := range tc.targets {
|
|
||||||
args = append(args, "-capture="+t)
|
|
||||||
}
|
|
||||||
|
|
||||||
if code := cmd.Run(args); code != 0 {
|
|
||||||
t.Fatalf("should exit 0, got code: %d", code)
|
|
||||||
}
|
|
||||||
|
|
||||||
errOutput := ui.ErrorWriter.String()
|
|
||||||
if errOutput != "" {
|
|
||||||
t.Errorf("expected no error output, got %q", errOutput)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ensure the debug data was written
|
|
||||||
_, err := os.Stat(outputPath)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatalf("output path should exist: %s", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ensure the captured static files exist
|
|
||||||
for _, f := range tc.files {
|
|
||||||
path := fmt.Sprintf("%s/%s", outputPath, f)
|
|
||||||
// Glob ignores file system errors
|
|
||||||
fs, _ := filepath.Glob(path)
|
|
||||||
if len(fs) <= 0 {
|
|
||||||
t.Fatalf("%s: output data should exist for %s", name, f)
|
|
||||||
}
|
|
||||||
for _, logFile := range fs {
|
|
||||||
content, err := ioutil.ReadFile(logFile)
|
|
||||||
require.NoError(t, err)
|
|
||||||
scanner := bufio.NewScanner(strings.NewReader(string(content)))
|
|
||||||
for scanner.Scan() {
|
|
||||||
logLine := scanner.Text()
|
|
||||||
if !validateLogLine([]byte(logLine)) {
|
|
||||||
t.Fatalf("%s: log line is not valid %s", name, logLine)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ensure any excluded files do not exist
|
|
||||||
for _, f := range tc.excludedFiles {
|
|
||||||
path := fmt.Sprintf("%s/%s", outputPath, f)
|
|
||||||
// Glob ignores file system errors
|
|
||||||
fs, _ := filepath.Glob(path)
|
|
||||||
if len(fs) > 0 {
|
|
||||||
t.Fatalf("%s: output data should not exist for %s", name, f)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func validateLogLine(content []byte) bool {
|
func validateLogLine(content []byte) bool {
|
||||||
fields := strings.SplitN(string(content), " ", 2)
|
fields := strings.SplitN(string(content), " ", 2)
|
||||||
if len(fields) != 2 {
|
if len(fields) != 2 {
|
||||||
|
@ -466,62 +437,6 @@ func validateLogLine(content []byte) bool {
|
||||||
return valid
|
return valid
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDebugCommand_ProfilesExist(t *testing.T) {
|
|
||||||
if testing.Short() {
|
|
||||||
t.Skip("too slow for testing.Short")
|
|
||||||
}
|
|
||||||
|
|
||||||
testDir := testutil.TempDir(t, "debug")
|
|
||||||
|
|
||||||
a := agent.NewTestAgent(t, `
|
|
||||||
enable_debug = true
|
|
||||||
`)
|
|
||||||
defer a.Shutdown()
|
|
||||||
testrpc.WaitForLeader(t, a.RPC, "dc1")
|
|
||||||
|
|
||||||
ui := cli.NewMockUi()
|
|
||||||
cmd := New(ui)
|
|
||||||
cmd.validateTiming = false
|
|
||||||
|
|
||||||
outputPath := fmt.Sprintf("%s/debug", testDir)
|
|
||||||
println(outputPath)
|
|
||||||
args := []string{
|
|
||||||
"-http-addr=" + a.HTTPAddr(),
|
|
||||||
"-output=" + outputPath,
|
|
||||||
// CPU profile has a minimum of 1s
|
|
||||||
"-archive=false",
|
|
||||||
"-duration=2s",
|
|
||||||
"-interval=1s",
|
|
||||||
"-capture=pprof",
|
|
||||||
}
|
|
||||||
|
|
||||||
if code := cmd.Run(args); code != 0 {
|
|
||||||
t.Fatalf("should exit 0, got code: %d", code)
|
|
||||||
}
|
|
||||||
|
|
||||||
profiles := []string{"heap.prof", "profile.prof", "goroutine.prof", "trace.out"}
|
|
||||||
// Glob ignores file system errors
|
|
||||||
for _, v := range profiles {
|
|
||||||
fs, _ := filepath.Glob(fmt.Sprintf("%s/*/%s", outputPath, v))
|
|
||||||
if len(fs) < 1 {
|
|
||||||
t.Errorf("output data should exist for %s", v)
|
|
||||||
}
|
|
||||||
for _, f := range fs {
|
|
||||||
if !strings.Contains(f, "trace.out") {
|
|
||||||
content, err := ioutil.ReadFile(f)
|
|
||||||
require.NoError(t, err)
|
|
||||||
_, err = profile.ParseData(content)
|
|
||||||
require.NoError(t, err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
errOutput := ui.ErrorWriter.String()
|
|
||||||
if errOutput != "" {
|
|
||||||
t.Errorf("expected no error output, got %s", errOutput)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDebugCommand_Prepare_ValidateTiming(t *testing.T) {
|
func TestDebugCommand_Prepare_ValidateTiming(t *testing.T) {
|
||||||
cases := map[string]struct {
|
cases := map[string]struct {
|
||||||
duration string
|
duration string
|
||||||
|
|
Loading…
Reference in New Issue