ed14061578
* Work on raft backend * Add logstore locally * Add encryptor and unsealable interfaces * Add clustering support to raft * Remove client and handler * Bootstrap raft on init * Cleanup raft logic a bit * More raft work * Work on TLS config * More work on bootstrapping * Fix build * More work on bootstrapping * More bootstrapping work * fix build * Remove consul dep * Fix build * merged oss/master into raft-storage * Work on bootstrapping * Get bootstrapping to work * Clean up FMS and node-id * Update local node ID logic * Cleanup node-id change * Work on snapshotting * Raft: Add remove peer API (#906) * Add remove peer API * Add some comments * Fix existing snapshotting (#909) * Raft get peers API (#912) * Read raft configuration * address review feedback * Use the Leadership Transfer API to step-down the active node (#918) * Raft join and unseal using Shamir keys (#917) * Raft join using shamir * Store AEAD instead of master key * Split the raft join process to answer the challenge after a successful unseal * get the follower to standby state * Make unseal work * minor changes * Some input checks * reuse the shamir seal access instead of new default seal access * refactor joinRaftSendAnswer function * Synchronously send answer in auto-unseal case * Address review feedback * Raft snapshots (#910) * Fix existing snapshotting * implement the noop snapshotting * Add comments and switch log libraries * add some snapshot tests * add snapshot test file * add TODO * More work on raft snapshotting * progress on the ConfigStore strategy * Don't use two buckets * Update the snapshot store logic to hide the file logic * Add more backend tests * Cleanup code a bit * [WIP] Raft recovery (#938) * Add recovery functionality * remove fmt.Printfs * Fix a few fsm bugs * Add max size value for raft backend (#942) * Add max size value for raft backend * Include physical.ErrValueTooLarge in the message * Raft snapshot Take/Restore API (#926) * Inital work on raft snapshot APIs * Always redirect snapshot install/download requests * More work on the snapshot APIs * Cleanup code a bit * On restore handle special cases * Use the seal to encrypt the sha sum file * Add sealer mechanism and fix some bugs * Call restore while state lock is held * Send restore cb trigger through raft log * Make error messages nicer * Add test helpers * Add snapshot test * Add shamir unseal test * Add more raft snapshot API tests * Fix locking * Change working to initalize * Add underlying raw object to test cluster core * Move leaderUUID to core * Add raft TLS rotation logic (#950) * Add TLS rotation logic * Cleanup logic a bit * Add/Remove from follower state on add/remove peer * add comments * Update more comments * Update request_forwarding_service.proto * Make sure we populate all nodes in the followerstate obj * Update times * Apply review feedback * Add more raft config setting (#947) * Add performance config setting * Add more config options and fix tests * Test Raft Recovery (#944) * Test raft recovery * Leave out a node during recovery * remove unused struct * Update physical/raft/snapshot_test.go * Update physical/raft/snapshot_test.go * fix vendoring * Switch to new raft interface * Remove unused files * Switch a gogo -> proto instance * Remove unneeded vault dep in go.sum * Update helper/testhelpers/testhelpers.go Co-Authored-By: Calvin Leung Huang <cleung2010@gmail.com> * Update vault/cluster/cluster.go * track active key within the keyring itself (#6915) * track active key within the keyring itself * lookup and store using the active key ID * update docstring * minor refactor * Small text fixes (#6912) * Update physical/raft/raft.go Co-Authored-By: Calvin Leung Huang <cleung2010@gmail.com> * review feedback * Move raft logical system into separate file * Update help text a bit * Enforce cluster addr is set and use it for raft bootstrapping * Fix tests * fix http test panic * Pull in latest raft-snapshot library * Add comment
279 lines
7.1 KiB
Go
279 lines
7.1 KiB
Go
package metrics
|
|
|
|
import (
|
|
"runtime"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/hashicorp/go-immutable-radix"
|
|
)
|
|
|
|
type Label struct {
|
|
Name string
|
|
Value string
|
|
}
|
|
|
|
func (m *Metrics) SetGauge(key []string, val float32) {
|
|
m.SetGaugeWithLabels(key, val, nil)
|
|
}
|
|
|
|
func (m *Metrics) SetGaugeWithLabels(key []string, val float32, labels []Label) {
|
|
if m.HostName != "" {
|
|
if m.EnableHostnameLabel {
|
|
labels = append(labels, Label{"host", m.HostName})
|
|
} else if m.EnableHostname {
|
|
key = insert(0, m.HostName, key)
|
|
}
|
|
}
|
|
if m.EnableTypePrefix {
|
|
key = insert(0, "gauge", key)
|
|
}
|
|
if m.ServiceName != "" {
|
|
if m.EnableServiceLabel {
|
|
labels = append(labels, Label{"service", m.ServiceName})
|
|
} else {
|
|
key = insert(0, m.ServiceName, key)
|
|
}
|
|
}
|
|
allowed, labelsFiltered := m.allowMetric(key, labels)
|
|
if !allowed {
|
|
return
|
|
}
|
|
m.sink.SetGaugeWithLabels(key, val, labelsFiltered)
|
|
}
|
|
|
|
func (m *Metrics) EmitKey(key []string, val float32) {
|
|
if m.EnableTypePrefix {
|
|
key = insert(0, "kv", key)
|
|
}
|
|
if m.ServiceName != "" {
|
|
key = insert(0, m.ServiceName, key)
|
|
}
|
|
allowed, _ := m.allowMetric(key, nil)
|
|
if !allowed {
|
|
return
|
|
}
|
|
m.sink.EmitKey(key, val)
|
|
}
|
|
|
|
func (m *Metrics) IncrCounter(key []string, val float32) {
|
|
m.IncrCounterWithLabels(key, val, nil)
|
|
}
|
|
|
|
func (m *Metrics) IncrCounterWithLabels(key []string, val float32, labels []Label) {
|
|
if m.HostName != "" && m.EnableHostnameLabel {
|
|
labels = append(labels, Label{"host", m.HostName})
|
|
}
|
|
if m.EnableTypePrefix {
|
|
key = insert(0, "counter", key)
|
|
}
|
|
if m.ServiceName != "" {
|
|
if m.EnableServiceLabel {
|
|
labels = append(labels, Label{"service", m.ServiceName})
|
|
} else {
|
|
key = insert(0, m.ServiceName, key)
|
|
}
|
|
}
|
|
allowed, labelsFiltered := m.allowMetric(key, labels)
|
|
if !allowed {
|
|
return
|
|
}
|
|
m.sink.IncrCounterWithLabels(key, val, labelsFiltered)
|
|
}
|
|
|
|
func (m *Metrics) AddSample(key []string, val float32) {
|
|
m.AddSampleWithLabels(key, val, nil)
|
|
}
|
|
|
|
func (m *Metrics) AddSampleWithLabels(key []string, val float32, labels []Label) {
|
|
if m.HostName != "" && m.EnableHostnameLabel {
|
|
labels = append(labels, Label{"host", m.HostName})
|
|
}
|
|
if m.EnableTypePrefix {
|
|
key = insert(0, "sample", key)
|
|
}
|
|
if m.ServiceName != "" {
|
|
if m.EnableServiceLabel {
|
|
labels = append(labels, Label{"service", m.ServiceName})
|
|
} else {
|
|
key = insert(0, m.ServiceName, key)
|
|
}
|
|
}
|
|
allowed, labelsFiltered := m.allowMetric(key, labels)
|
|
if !allowed {
|
|
return
|
|
}
|
|
m.sink.AddSampleWithLabels(key, val, labelsFiltered)
|
|
}
|
|
|
|
func (m *Metrics) MeasureSince(key []string, start time.Time) {
|
|
m.MeasureSinceWithLabels(key, start, nil)
|
|
}
|
|
|
|
func (m *Metrics) MeasureSinceWithLabels(key []string, start time.Time, labels []Label) {
|
|
if m.HostName != "" && m.EnableHostnameLabel {
|
|
labels = append(labels, Label{"host", m.HostName})
|
|
}
|
|
if m.EnableTypePrefix {
|
|
key = insert(0, "timer", key)
|
|
}
|
|
if m.ServiceName != "" {
|
|
if m.EnableServiceLabel {
|
|
labels = append(labels, Label{"service", m.ServiceName})
|
|
} else {
|
|
key = insert(0, m.ServiceName, key)
|
|
}
|
|
}
|
|
allowed, labelsFiltered := m.allowMetric(key, labels)
|
|
if !allowed {
|
|
return
|
|
}
|
|
now := time.Now()
|
|
elapsed := now.Sub(start)
|
|
msec := float32(elapsed.Nanoseconds()) / float32(m.TimerGranularity)
|
|
m.sink.AddSampleWithLabels(key, msec, labelsFiltered)
|
|
}
|
|
|
|
// UpdateFilter overwrites the existing filter with the given rules.
|
|
func (m *Metrics) UpdateFilter(allow, block []string) {
|
|
m.UpdateFilterAndLabels(allow, block, m.AllowedLabels, m.BlockedLabels)
|
|
}
|
|
|
|
// UpdateFilterAndLabels overwrites the existing filter with the given rules.
|
|
func (m *Metrics) UpdateFilterAndLabels(allow, block, allowedLabels, blockedLabels []string) {
|
|
m.filterLock.Lock()
|
|
defer m.filterLock.Unlock()
|
|
|
|
m.AllowedPrefixes = allow
|
|
m.BlockedPrefixes = block
|
|
|
|
if allowedLabels == nil {
|
|
// Having a white list means we take only elements from it
|
|
m.allowedLabels = nil
|
|
} else {
|
|
m.allowedLabels = make(map[string]bool)
|
|
for _, v := range allowedLabels {
|
|
m.allowedLabels[v] = true
|
|
}
|
|
}
|
|
m.blockedLabels = make(map[string]bool)
|
|
for _, v := range blockedLabels {
|
|
m.blockedLabels[v] = true
|
|
}
|
|
m.AllowedLabels = allowedLabels
|
|
m.BlockedLabels = blockedLabels
|
|
|
|
m.filter = iradix.New()
|
|
for _, prefix := range m.AllowedPrefixes {
|
|
m.filter, _, _ = m.filter.Insert([]byte(prefix), true)
|
|
}
|
|
for _, prefix := range m.BlockedPrefixes {
|
|
m.filter, _, _ = m.filter.Insert([]byte(prefix), false)
|
|
}
|
|
}
|
|
|
|
// labelIsAllowed return true if a should be included in metric
|
|
// the caller should lock m.filterLock while calling this method
|
|
func (m *Metrics) labelIsAllowed(label *Label) bool {
|
|
labelName := (*label).Name
|
|
if m.blockedLabels != nil {
|
|
_, ok := m.blockedLabels[labelName]
|
|
if ok {
|
|
// If present, let's remove this label
|
|
return false
|
|
}
|
|
}
|
|
if m.allowedLabels != nil {
|
|
_, ok := m.allowedLabels[labelName]
|
|
return ok
|
|
}
|
|
// Allow by default
|
|
return true
|
|
}
|
|
|
|
// filterLabels return only allowed labels
|
|
// the caller should lock m.filterLock while calling this method
|
|
func (m *Metrics) filterLabels(labels []Label) []Label {
|
|
if labels == nil {
|
|
return nil
|
|
}
|
|
toReturn := []Label{}
|
|
for _, label := range labels {
|
|
if m.labelIsAllowed(&label) {
|
|
toReturn = append(toReturn, label)
|
|
}
|
|
}
|
|
return toReturn
|
|
}
|
|
|
|
// Returns whether the metric should be allowed based on configured prefix filters
|
|
// Also return the applicable labels
|
|
func (m *Metrics) allowMetric(key []string, labels []Label) (bool, []Label) {
|
|
m.filterLock.RLock()
|
|
defer m.filterLock.RUnlock()
|
|
|
|
if m.filter == nil || m.filter.Len() == 0 {
|
|
return m.Config.FilterDefault, m.filterLabels(labels)
|
|
}
|
|
|
|
_, allowed, ok := m.filter.Root().LongestPrefix([]byte(strings.Join(key, ".")))
|
|
if !ok {
|
|
return m.Config.FilterDefault, m.filterLabels(labels)
|
|
}
|
|
|
|
return allowed.(bool), m.filterLabels(labels)
|
|
}
|
|
|
|
// Periodically collects runtime stats to publish
|
|
func (m *Metrics) collectStats() {
|
|
for {
|
|
time.Sleep(m.ProfileInterval)
|
|
m.emitRuntimeStats()
|
|
}
|
|
}
|
|
|
|
// Emits various runtime statsitics
|
|
func (m *Metrics) emitRuntimeStats() {
|
|
// Export number of Goroutines
|
|
numRoutines := runtime.NumGoroutine()
|
|
m.SetGauge([]string{"runtime", "num_goroutines"}, float32(numRoutines))
|
|
|
|
// Export memory stats
|
|
var stats runtime.MemStats
|
|
runtime.ReadMemStats(&stats)
|
|
m.SetGauge([]string{"runtime", "alloc_bytes"}, float32(stats.Alloc))
|
|
m.SetGauge([]string{"runtime", "sys_bytes"}, float32(stats.Sys))
|
|
m.SetGauge([]string{"runtime", "malloc_count"}, float32(stats.Mallocs))
|
|
m.SetGauge([]string{"runtime", "free_count"}, float32(stats.Frees))
|
|
m.SetGauge([]string{"runtime", "heap_objects"}, float32(stats.HeapObjects))
|
|
m.SetGauge([]string{"runtime", "total_gc_pause_ns"}, float32(stats.PauseTotalNs))
|
|
m.SetGauge([]string{"runtime", "total_gc_runs"}, float32(stats.NumGC))
|
|
|
|
// Export info about the last few GC runs
|
|
num := stats.NumGC
|
|
|
|
// Handle wrap around
|
|
if num < m.lastNumGC {
|
|
m.lastNumGC = 0
|
|
}
|
|
|
|
// Ensure we don't scan more than 256
|
|
if num-m.lastNumGC >= 256 {
|
|
m.lastNumGC = num - 255
|
|
}
|
|
|
|
for i := m.lastNumGC; i < num; i++ {
|
|
pause := stats.PauseNs[i%256]
|
|
m.AddSample([]string{"runtime", "gc_pause_ns"}, float32(pause))
|
|
}
|
|
m.lastNumGC = num
|
|
}
|
|
|
|
// Inserts a string value at an index into the slice
|
|
func insert(i int, v string, s []string) []string {
|
|
s = append(s, "")
|
|
copy(s[i+1:], s[i:])
|
|
s[i] = v
|
|
return s
|
|
}
|