820af27171
Saves a tiny bit of cpu and some IO. Sadly doesn't prevent all IO on duplicate writes as the transactions are still created and committed. $ go test -bench=. -benchmem goos: linux goarch: amd64 pkg: github.com/hashicorp/nomad/helper/boltdd BenchmarkWriteDeduplication_On-4 500 4059591 ns/op 23736 B/op 56 allocs/op BenchmarkWriteDeduplication_Off-4 300 4115319 ns/op 25942 B/op 55 allocs/op
345 lines
8.1 KiB
Go
345 lines
8.1 KiB
Go
package boltdd
|
|
|
|
import (
|
|
"bytes"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"os"
|
|
"path/filepath"
|
|
"testing"
|
|
|
|
"github.com/hashicorp/nomad/nomad/mock"
|
|
"github.com/hashicorp/nomad/nomad/structs"
|
|
"github.com/stretchr/testify/require"
|
|
"github.com/ugorji/go/codec"
|
|
)
|
|
|
|
type testingT interface {
|
|
Fatalf(format string, args ...interface{})
|
|
Logf(format string, args ...interface{})
|
|
}
|
|
|
|
func setupBoltDB(t testingT) (*DB, func()) {
|
|
dir, err := ioutil.TempDir("", "nomadtest_")
|
|
if err != nil {
|
|
t.Fatalf("error creating tempdir: %v", err)
|
|
}
|
|
|
|
cleanup := func() {
|
|
if err := os.RemoveAll(dir); err != nil {
|
|
t.Logf("error removing test dir: %v", err)
|
|
}
|
|
}
|
|
|
|
dbFilename := filepath.Join(dir, "nomadtest.db")
|
|
db, err := Open(dbFilename, 0600, nil)
|
|
if err != nil {
|
|
cleanup()
|
|
t.Fatalf("error creating boltdb: %v", err)
|
|
}
|
|
|
|
return db, func() {
|
|
db.Close()
|
|
cleanup()
|
|
}
|
|
}
|
|
|
|
func TestDB_Open(t *testing.T) {
|
|
t.Parallel()
|
|
require := require.New(t)
|
|
|
|
db, cleanup := setupBoltDB(t)
|
|
defer cleanup()
|
|
|
|
require.Equal(0, db.BoltDB().Stats().TxStats.Write)
|
|
}
|
|
|
|
func TestBucket_Create(t *testing.T) {
|
|
t.Parallel()
|
|
require := require.New(t)
|
|
|
|
db, cleanup := setupBoltDB(t)
|
|
defer cleanup()
|
|
|
|
name := []byte("create_test")
|
|
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
// Trying to get a nonexistent bucket should return nil
|
|
require.Nil(tx.Bucket(name))
|
|
|
|
// Creating a nonexistent bucket should work
|
|
b, err := tx.CreateBucket(name)
|
|
require.NoError(err)
|
|
require.NotNil(b)
|
|
|
|
// Recreating a bucket that exists should fail
|
|
b, err = tx.CreateBucket(name)
|
|
require.Error(err)
|
|
require.Nil(b)
|
|
|
|
// get or create should work
|
|
b, err = tx.CreateBucketIfNotExists(name)
|
|
require.NoError(err)
|
|
require.NotNil(b)
|
|
return nil
|
|
}))
|
|
|
|
// Bucket should be visible
|
|
require.NoError(db.View(func(tx *Tx) error {
|
|
require.NotNil(tx.Bucket(name))
|
|
return nil
|
|
}))
|
|
}
|
|
|
|
func TestBucket_DedupeWrites(t *testing.T) {
|
|
t.Parallel()
|
|
require := require.New(t)
|
|
|
|
db, cleanup := setupBoltDB(t)
|
|
defer cleanup()
|
|
|
|
bname := []byte("dedupewrites_test")
|
|
k1name := []byte("k1")
|
|
k2name := []byte("k2")
|
|
|
|
// Put 2 keys
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
b, err := tx.CreateBucket(bname)
|
|
require.NoError(err)
|
|
|
|
require.NoError(b.Put(k1name, k1name))
|
|
require.NoError(b.Put(k2name, k2name))
|
|
return nil
|
|
}))
|
|
|
|
// Assert there was at least 1 write
|
|
origWrites := db.BoltDB().Stats().TxStats.Write
|
|
require.NotZero(origWrites)
|
|
|
|
// Write the same values again and expect no new writes
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
b := tx.Bucket(bname)
|
|
require.NoError(b.Put(k1name, k1name))
|
|
require.NoError(b.Put(k2name, k2name))
|
|
return nil
|
|
}))
|
|
|
|
putWrites := db.BoltDB().Stats().TxStats.Write
|
|
|
|
// Unforunately every committed transaction causes two writes, so this
|
|
// only saves 1 write operation
|
|
require.Equal(origWrites+2, putWrites)
|
|
|
|
// Write new values and assert more writes took place
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
b := tx.Bucket(bname)
|
|
require.NoError(b.Put(k1name, []byte("newval1")))
|
|
require.NoError(b.Put(k2name, []byte("newval2")))
|
|
return nil
|
|
}))
|
|
|
|
putWrites2 := db.BoltDB().Stats().TxStats.Write
|
|
|
|
// Expect 3 additional writes: 2 for the transaction and one for the
|
|
// dirty page
|
|
require.Equal(putWrites+3, putWrites2)
|
|
}
|
|
|
|
func TestBucket_Delete(t *testing.T) {
|
|
t.Parallel()
|
|
require := require.New(t)
|
|
|
|
db, cleanup := setupBoltDB(t)
|
|
defer cleanup()
|
|
|
|
parentName := []byte("delete_test")
|
|
parentKey := []byte("parent_key")
|
|
childName := []byte("child")
|
|
childKey := []byte("child_key")
|
|
grandchildName1 := []byte("grandchild1")
|
|
grandchildKey1 := []byte("grandchild_key1")
|
|
grandchildName2 := []byte("grandchild2")
|
|
grandchildKey2 := []byte("grandchild_key2")
|
|
|
|
// Create a parent bucket with 1 child and 2 grandchildren
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
pb, err := tx.CreateBucket(parentName)
|
|
require.NoError(err)
|
|
|
|
require.NoError(pb.Put(parentKey, parentKey))
|
|
|
|
child, err := pb.CreateBucket(childName)
|
|
require.NoError(err)
|
|
|
|
require.NoError(child.Put(childKey, childKey))
|
|
|
|
grandchild1, err := child.CreateBucket(grandchildName1)
|
|
require.NoError(err)
|
|
|
|
require.NoError(grandchild1.Put(grandchildKey1, grandchildKey1))
|
|
|
|
grandchild2, err := child.CreateBucket(grandchildName2)
|
|
require.NoError(err)
|
|
|
|
require.NoError(grandchild2.Put(grandchildKey2, grandchildKey2))
|
|
return nil
|
|
}))
|
|
|
|
// Verify grandchild keys wrote
|
|
require.NoError(db.View(func(tx *Tx) error {
|
|
grandchild1 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName1)
|
|
var v1 []byte
|
|
grandchild1.Get(grandchildKey1, &v1)
|
|
require.Equal(grandchildKey1, v1)
|
|
|
|
grandchild2 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName2)
|
|
var v2 []byte
|
|
grandchild2.Get(grandchildKey2, &v2)
|
|
require.Equal(grandchildKey2, v2)
|
|
return nil
|
|
}))
|
|
|
|
// Delete grandchildKey1 and grandchild2
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
child := tx.Bucket(parentName).Bucket(childName)
|
|
|
|
require.NoError(child.DeleteBucket(grandchildName2))
|
|
|
|
grandchild1 := child.Bucket(grandchildName1)
|
|
require.NoError(grandchild1.Delete(grandchildKey1))
|
|
return nil
|
|
}))
|
|
|
|
// Ensure grandchild2 alone was deleted
|
|
require.NoError(db.View(func(tx *Tx) error {
|
|
grandchild1 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName1)
|
|
var v1 []byte
|
|
grandchild1.Get(grandchildKey1, &v1)
|
|
require.Equal(([]byte)(nil), v1)
|
|
|
|
grandchild2 := tx.Bucket(parentName).Bucket(childName).Bucket(grandchildName2)
|
|
require.Nil(grandchild2)
|
|
return nil
|
|
}))
|
|
|
|
// Deleting child bucket should delete grandchild1 as well
|
|
require.NoError(db.Update(func(tx *Tx) error {
|
|
parent := tx.Bucket(parentName)
|
|
require.NoError(parent.DeleteBucket(childName))
|
|
|
|
// Recreate child bucket and ensure childKey and grandchild are gone
|
|
child, err := parent.CreateBucket(childName)
|
|
require.NoError(err)
|
|
|
|
var v []byte
|
|
require.Error(child.Get(childKey, &v))
|
|
require.Equal(([]byte)(nil), v)
|
|
|
|
require.Nil(child.Bucket(grandchildName1))
|
|
|
|
// Rewrite childKey1 to make sure it doesn't get dedupe incorrectly
|
|
require.NoError(child.Put(childKey, childKey))
|
|
return nil
|
|
}))
|
|
|
|
// Ensure childKey1 was rewritten and not deduped incorrectly
|
|
require.NoError(db.View(func(tx *Tx) error {
|
|
var v []byte
|
|
require.NoError(tx.Bucket(parentName).Bucket(childName).Get(childKey, &v))
|
|
require.Equal(childKey, v)
|
|
return nil
|
|
}))
|
|
}
|
|
|
|
func BenchmarkWriteDeduplication_On(b *testing.B) {
|
|
db, cleanup := setupBoltDB(b)
|
|
defer cleanup()
|
|
|
|
bucketName := []byte("allocations")
|
|
alloc := mock.Alloc()
|
|
allocID := []byte(alloc.ID)
|
|
|
|
err := db.Update(func(tx *Tx) error {
|
|
allocs, err := tx.CreateBucket(bucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return allocs.Put(allocID, alloc)
|
|
})
|
|
|
|
if err != nil {
|
|
b.Fatalf("error setting up: %v", err)
|
|
}
|
|
|
|
b.ResetTimer()
|
|
for i := 0; i < b.N; i++ {
|
|
err := db.Update(func(tx *Tx) error {
|
|
return tx.Bucket(bucketName).Put(allocID, alloc)
|
|
})
|
|
|
|
if err != nil {
|
|
b.Fatalf("error at runtime: %v", err)
|
|
}
|
|
}
|
|
}
|
|
|
|
func BenchmarkWriteDeduplication_Off(b *testing.B) {
|
|
dir, err := ioutil.TempDir("", "nomadtest_")
|
|
if err != nil {
|
|
b.Fatalf("error creating tempdir: %v", err)
|
|
}
|
|
|
|
defer func() {
|
|
if err := os.RemoveAll(dir); err != nil {
|
|
b.Logf("error removing test dir: %v", err)
|
|
}
|
|
}()
|
|
|
|
dbFilename := filepath.Join(dir, "nomadtest.db")
|
|
db, err := Open(dbFilename, 0600, nil)
|
|
if err != nil {
|
|
b.Fatalf("error creating boltdb: %v", err)
|
|
}
|
|
|
|
defer db.Close()
|
|
|
|
bucketName := []byte("allocations")
|
|
alloc := mock.Alloc()
|
|
allocID := []byte(alloc.ID)
|
|
|
|
err = db.Update(func(tx *Tx) error {
|
|
allocs, err := tx.CreateBucket(bucketName)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
var buf bytes.Buffer
|
|
if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(alloc); err != nil {
|
|
return fmt.Errorf("failed to encode passed object: %v", err)
|
|
}
|
|
|
|
return allocs.Put(allocID, buf)
|
|
})
|
|
|
|
if err != nil {
|
|
b.Fatalf("error setting up: %v", err)
|
|
}
|
|
|
|
b.ResetTimer()
|
|
for i := 0; i < b.N; i++ {
|
|
err := db.Update(func(tx *Tx) error {
|
|
var buf bytes.Buffer
|
|
if err := codec.NewEncoder(&buf, structs.MsgpackHandle).Encode(alloc); err != nil {
|
|
return fmt.Errorf("failed to encode passed object: %v", err)
|
|
}
|
|
|
|
return tx.Bucket(bucketName).Put(allocID, buf)
|
|
})
|
|
|
|
if err != nil {
|
|
b.Fatalf("error at runtime: %v", err)
|
|
}
|
|
}
|
|
}
|