mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-30 13:41:46 +00:00
cc6f323705
Summary: compensate file sizes in compaction picking so files with range tombstones are preferred, such that they get compacted down earlier as they tend to delete a lot of data. This PR adds a `compensated_range_deletion_size` field in FileMeta that is computed during Flush/Compaction and persisted in MANIFEST. This value is added to `compensated_file_size` which will be used for compaction picking. Currently, for a file in level L, `compensated_range_deletion_size` is set to the estimated bytes deleted by range tombstone of this file in all levels > L. This helps to reduce space amp when data in older levels are covered by range tombstones in level L. Pull Request resolved: https://github.com/facebook/rocksdb/pull/10734 Test Plan: - Added unit tests. - benchmark to check if the above definition `compensated_range_deletion_size` is reducing space amp as intended, without affecting write amp too much. The experiment set up favorable for this optimization: large range tombstone issued infrequently. Command used: ``` ./db_bench -benchmarks=fillrandom,waitforcompaction,stats,levelstats -use_existing_db=false -avoid_flush_during_recovery=true -write_buffer_size=33554432 -level_compaction_dynamic_level_bytes=true -max_background_jobs=8 -max_bytes_for_level_base=134217728 -target_file_size_base=33554432 -writes_per_range_tombstone=500000 -range_tombstone_width=5000000 -num=50000000 -benchmark_write_rate_limit=8388608 -threads=16 -duration=1800 --max_num_range_tombstones=1000000000 ``` In this experiment, each thread wrote 16 range tombstones over the duration of 30 minutes, each range tombstone has width 5M that is the 10% of the key space width. Results shows this PR generates a smaller DB size. Compaction stats from this PR: ``` Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ L0 2/0 31.54 MB 0.5 0.0 0.0 0.0 8.4 8.4 0.0 1.0 0.0 63.4 135.56 110.94 544 0.249 0 0 0.0 0.0 L4 3/0 96.55 MB 0.8 18.5 6.7 11.8 18.4 6.6 0.0 2.7 65.3 64.9 290.08 284.03 108 2.686 284M 1957K 0.0 0.0 L5 15/0 404.41 MB 1.0 19.1 7.7 11.4 18.8 7.4 0.3 2.5 66.6 65.7 292.93 285.34 220 1.332 293M 3808K 0.0 0.0 L6 143/0 4.12 GB 0.0 45.0 7.5 37.5 41.6 4.1 0.0 5.5 71.2 65.9 647.00 632.66 251 2.578 739M 47M 0.0 0.0 Sum 163/0 4.64 GB 0.0 82.6 21.9 60.7 87.2 26.5 0.3 10.4 61.9 65.4 1365.58 1312.97 1123 1.216 1318M 52M 0.0 0.0 ``` Compaction stats from main: ``` Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ L0 0/0 0.00 KB 0.0 0.0 0.0 0.0 8.4 8.4 0.0 1.0 0.0 60.5 142.12 115.89 569 0.250 0 0 0.0 0.0 L4 3/0 85.68 MB 1.0 17.7 6.8 10.9 17.6 6.7 0.0 2.6 62.7 62.3 289.05 281.79 112 2.581 272M 2309K 0.0 0.0 L5 11/0 293.73 MB 1.0 18.8 7.5 11.2 18.5 7.2 0.5 2.5 64.9 63.9 296.07 288.50 220 1.346 288M 4365K 0.0 0.0 L6 130/0 3.94 GB 0.0 51.5 7.6 43.9 47.9 3.9 0.0 6.3 67.2 62.4 784.95 765.92 258 3.042 848M 51M 0.0 0.0 Sum 144/0 4.31 GB 0.0 88.0 21.9 66.0 92.3 26.3 0.5 11.0 59.6 62.5 1512.19 1452.09 1159 1.305 1409M 58M 0.0 0.0``` Reviewed By: ajkr Differential Revision: D39834713 Pulled By: cbi42 fbshipit-source-id: fe9341040b8704a8fbb10cad5cf5c43e962c7e6b
733 lines
24 KiB
C++
733 lines
24 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#include "db/version_edit.h"
|
|
|
|
#include "db/blob/blob_index.h"
|
|
#include "rocksdb/advanced_options.h"
|
|
#include "table/unique_id_impl.h"
|
|
#include "test_util/sync_point.h"
|
|
#include "test_util/testharness.h"
|
|
#include "test_util/testutil.h"
|
|
#include "util/coding.h"
|
|
#include "util/string_util.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
static void TestEncodeDecode(const VersionEdit& edit) {
|
|
std::string encoded, encoded2;
|
|
edit.EncodeTo(&encoded);
|
|
VersionEdit parsed;
|
|
Status s = parsed.DecodeFrom(encoded);
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
parsed.EncodeTo(&encoded2);
|
|
ASSERT_EQ(encoded, encoded2);
|
|
}
|
|
|
|
class VersionEditTest : public testing::Test {};
|
|
|
|
TEST_F(VersionEditTest, EncodeDecode) {
|
|
static const uint64_t kBig = 1ull << 50;
|
|
static const uint32_t kBig32Bit = 1ull << 30;
|
|
|
|
VersionEdit edit;
|
|
for (int i = 0; i < 4; i++) {
|
|
TestEncodeDecode(edit);
|
|
edit.AddFile(3, kBig + 300 + i, kBig32Bit + 400 + i, 0,
|
|
InternalKey("foo", kBig + 500 + i, kTypeValue),
|
|
InternalKey("zoo", kBig + 600 + i, kTypeDeletion),
|
|
kBig + 500 + i, kBig + 600 + i, false, Temperature::kUnknown,
|
|
kInvalidBlobFileNumber, 888, 678,
|
|
kBig + 300 + i /* epoch_number */, "234", "crc32c",
|
|
kNullUniqueId64x2, 0);
|
|
edit.DeleteFile(4, kBig + 700 + i);
|
|
}
|
|
|
|
edit.SetComparatorName("foo");
|
|
edit.SetLogNumber(kBig + 100);
|
|
edit.SetNextFile(kBig + 200);
|
|
edit.SetLastSequence(kBig + 1000);
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, EncodeDecodeNewFile4) {
|
|
static const uint64_t kBig = 1ull << 50;
|
|
|
|
VersionEdit edit;
|
|
edit.AddFile(3, 300, 3, 100, InternalKey("foo", kBig + 500, kTypeValue),
|
|
InternalKey("zoo", kBig + 600, kTypeDeletion), kBig + 500,
|
|
kBig + 600, true, Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
kUnknownOldestAncesterTime, kUnknownFileCreationTime,
|
|
300 /* epoch_number */, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
edit.AddFile(4, 301, 3, 100, InternalKey("foo", kBig + 501, kTypeValue),
|
|
InternalKey("zoo", kBig + 601, kTypeDeletion), kBig + 501,
|
|
kBig + 601, false, Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
kUnknownOldestAncesterTime, kUnknownFileCreationTime,
|
|
301 /* epoch_number */, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
edit.AddFile(5, 302, 0, 100, InternalKey("foo", kBig + 502, kTypeValue),
|
|
InternalKey("zoo", kBig + 602, kTypeDeletion), kBig + 502,
|
|
kBig + 602, true, Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
666, 888, 302 /* epoch_number */, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
edit.AddFile(5, 303, 0, 100, InternalKey("foo", kBig + 503, kTypeBlobIndex),
|
|
InternalKey("zoo", kBig + 603, kTypeBlobIndex), kBig + 503,
|
|
kBig + 603, true, Temperature::kUnknown, 1001,
|
|
kUnknownOldestAncesterTime, kUnknownFileCreationTime,
|
|
303 /* epoch_number */, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
|
|
edit.DeleteFile(4, 700);
|
|
|
|
edit.SetComparatorName("foo");
|
|
edit.SetLogNumber(kBig + 100);
|
|
edit.SetNextFile(kBig + 200);
|
|
edit.SetLastSequence(kBig + 1000);
|
|
TestEncodeDecode(edit);
|
|
|
|
std::string encoded, encoded2;
|
|
edit.EncodeTo(&encoded);
|
|
VersionEdit parsed;
|
|
Status s = parsed.DecodeFrom(encoded);
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
auto& new_files = parsed.GetNewFiles();
|
|
ASSERT_TRUE(new_files[0].second.marked_for_compaction);
|
|
ASSERT_TRUE(!new_files[1].second.marked_for_compaction);
|
|
ASSERT_TRUE(new_files[2].second.marked_for_compaction);
|
|
ASSERT_TRUE(new_files[3].second.marked_for_compaction);
|
|
ASSERT_EQ(3u, new_files[0].second.fd.GetPathId());
|
|
ASSERT_EQ(3u, new_files[1].second.fd.GetPathId());
|
|
ASSERT_EQ(0u, new_files[2].second.fd.GetPathId());
|
|
ASSERT_EQ(0u, new_files[3].second.fd.GetPathId());
|
|
ASSERT_EQ(kInvalidBlobFileNumber,
|
|
new_files[0].second.oldest_blob_file_number);
|
|
ASSERT_EQ(kInvalidBlobFileNumber,
|
|
new_files[1].second.oldest_blob_file_number);
|
|
ASSERT_EQ(kInvalidBlobFileNumber,
|
|
new_files[2].second.oldest_blob_file_number);
|
|
ASSERT_EQ(1001, new_files[3].second.oldest_blob_file_number);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, ForwardCompatibleNewFile4) {
|
|
static const uint64_t kBig = 1ull << 50;
|
|
VersionEdit edit;
|
|
edit.AddFile(3, 300, 3, 100, InternalKey("foo", kBig + 500, kTypeValue),
|
|
InternalKey("zoo", kBig + 600, kTypeDeletion), kBig + 500,
|
|
kBig + 600, true, Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
kUnknownOldestAncesterTime, kUnknownFileCreationTime,
|
|
300 /* epoch_number */, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
edit.AddFile(4, 301, 3, 100, InternalKey("foo", kBig + 501, kTypeValue),
|
|
InternalKey("zoo", kBig + 601, kTypeDeletion), kBig + 501,
|
|
kBig + 601, false, Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
686, 868, 301 /* epoch_number */, "234", "crc32c",
|
|
kNullUniqueId64x2, 0);
|
|
edit.DeleteFile(4, 700);
|
|
|
|
edit.SetComparatorName("foo");
|
|
edit.SetLogNumber(kBig + 100);
|
|
edit.SetNextFile(kBig + 200);
|
|
edit.SetLastSequence(kBig + 1000);
|
|
|
|
std::string encoded;
|
|
|
|
// Call back function to add extra customized builds.
|
|
bool first = true;
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
|
|
"VersionEdit::EncodeTo:NewFile4:CustomizeFields", [&](void* arg) {
|
|
std::string* str = reinterpret_cast<std::string*>(arg);
|
|
PutVarint32(str, 33);
|
|
const std::string str1 = "random_string";
|
|
PutLengthPrefixedSlice(str, str1);
|
|
if (first) {
|
|
first = false;
|
|
PutVarint32(str, 22);
|
|
const std::string str2 = "s";
|
|
PutLengthPrefixedSlice(str, str2);
|
|
}
|
|
});
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
|
edit.EncodeTo(&encoded);
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
VersionEdit parsed;
|
|
Status s = parsed.DecodeFrom(encoded);
|
|
ASSERT_TRUE(s.ok()) << s.ToString();
|
|
ASSERT_TRUE(!first);
|
|
auto& new_files = parsed.GetNewFiles();
|
|
ASSERT_TRUE(new_files[0].second.marked_for_compaction);
|
|
ASSERT_TRUE(!new_files[1].second.marked_for_compaction);
|
|
ASSERT_EQ(3u, new_files[0].second.fd.GetPathId());
|
|
ASSERT_EQ(3u, new_files[1].second.fd.GetPathId());
|
|
ASSERT_EQ(1u, parsed.GetDeletedFiles().size());
|
|
}
|
|
|
|
TEST_F(VersionEditTest, NewFile4NotSupportedField) {
|
|
static const uint64_t kBig = 1ull << 50;
|
|
VersionEdit edit;
|
|
edit.AddFile(3, 300, 3, 100, InternalKey("foo", kBig + 500, kTypeValue),
|
|
InternalKey("zoo", kBig + 600, kTypeDeletion), kBig + 500,
|
|
kBig + 600, true, Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
kUnknownOldestAncesterTime, kUnknownFileCreationTime,
|
|
300 /* epoch_number */, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
|
|
edit.SetComparatorName("foo");
|
|
edit.SetLogNumber(kBig + 100);
|
|
edit.SetNextFile(kBig + 200);
|
|
edit.SetLastSequence(kBig + 1000);
|
|
|
|
std::string encoded;
|
|
|
|
// Call back function to add extra customized builds.
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
|
|
"VersionEdit::EncodeTo:NewFile4:CustomizeFields", [&](void* arg) {
|
|
std::string* str = reinterpret_cast<std::string*>(arg);
|
|
const std::string str1 = "s";
|
|
PutLengthPrefixedSlice(str, str1);
|
|
});
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
|
edit.EncodeTo(&encoded);
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
VersionEdit parsed;
|
|
Status s = parsed.DecodeFrom(encoded);
|
|
ASSERT_NOK(s);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, EncodeEmptyFile) {
|
|
VersionEdit edit;
|
|
edit.AddFile(0, 0, 0, 0, InternalKey(), InternalKey(), 0, 0, false,
|
|
Temperature::kUnknown, kInvalidBlobFileNumber,
|
|
kUnknownOldestAncesterTime, kUnknownFileCreationTime,
|
|
1 /*epoch_number*/, kUnknownFileChecksum,
|
|
kUnknownFileChecksumFuncName, kNullUniqueId64x2, 0);
|
|
std::string buffer;
|
|
ASSERT_TRUE(!edit.EncodeTo(&buffer));
|
|
}
|
|
|
|
TEST_F(VersionEditTest, ColumnFamilyTest) {
|
|
VersionEdit edit;
|
|
edit.SetColumnFamily(2);
|
|
edit.AddColumnFamily("column_family");
|
|
edit.SetMaxColumnFamily(5);
|
|
TestEncodeDecode(edit);
|
|
|
|
edit.Clear();
|
|
edit.SetColumnFamily(3);
|
|
edit.DropColumnFamily();
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, MinLogNumberToKeep) {
|
|
VersionEdit edit;
|
|
edit.SetMinLogNumberToKeep(13);
|
|
TestEncodeDecode(edit);
|
|
|
|
edit.Clear();
|
|
edit.SetMinLogNumberToKeep(23);
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, AtomicGroupTest) {
|
|
VersionEdit edit;
|
|
edit.MarkAtomicGroup(1);
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, IgnorableField) {
|
|
VersionEdit ve;
|
|
std::string encoded;
|
|
|
|
// Size of ignorable field is too large
|
|
PutVarint32Varint64(&encoded, 2 /* kLogNumber */, 66);
|
|
// This is a customized ignorable tag
|
|
PutVarint32Varint64(&encoded,
|
|
0x2710 /* A field with kTagSafeIgnoreMask set */,
|
|
5 /* fieldlength 5 */);
|
|
encoded += "abc"; // Only fills 3 bytes,
|
|
ASSERT_NOK(ve.DecodeFrom(encoded));
|
|
|
|
encoded.clear();
|
|
// Error when seeing unidentified tag that is not ignorable
|
|
PutVarint32Varint64(&encoded, 2 /* kLogNumber */, 66);
|
|
// This is a customized ignorable tag
|
|
PutVarint32Varint64(&encoded, 666 /* A field with kTagSafeIgnoreMask unset */,
|
|
3 /* fieldlength 3 */);
|
|
encoded += "abc"; // Fill 3 bytes
|
|
PutVarint32Varint64(&encoded, 3 /* next file number */, 88);
|
|
ASSERT_NOK(ve.DecodeFrom(encoded));
|
|
|
|
// Safely ignore an identified but safely ignorable entry
|
|
encoded.clear();
|
|
PutVarint32Varint64(&encoded, 2 /* kLogNumber */, 66);
|
|
// This is a customized ignorable tag
|
|
PutVarint32Varint64(&encoded,
|
|
0x2710 /* A field with kTagSafeIgnoreMask set */,
|
|
3 /* fieldlength 3 */);
|
|
encoded += "abc"; // Fill 3 bytes
|
|
PutVarint32Varint64(&encoded, 3 /* kNextFileNumber */, 88);
|
|
|
|
ASSERT_OK(ve.DecodeFrom(encoded));
|
|
|
|
ASSERT_TRUE(ve.HasLogNumber());
|
|
ASSERT_TRUE(ve.HasNextFile());
|
|
ASSERT_EQ(66, ve.GetLogNumber());
|
|
ASSERT_EQ(88, ve.GetNextFile());
|
|
}
|
|
|
|
TEST_F(VersionEditTest, DbId) {
|
|
VersionEdit edit;
|
|
edit.SetDBId("ab34-cd12-435f-er00");
|
|
TestEncodeDecode(edit);
|
|
|
|
edit.Clear();
|
|
edit.SetDBId("34ba-cd12-435f-er01");
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, BlobFileAdditionAndGarbage) {
|
|
VersionEdit edit;
|
|
|
|
const std::string checksum_method_prefix = "Hash";
|
|
const std::string checksum_value_prefix = "Value";
|
|
|
|
for (uint64_t blob_file_number = 1; blob_file_number <= 10;
|
|
++blob_file_number) {
|
|
const uint64_t total_blob_count = blob_file_number << 10;
|
|
const uint64_t total_blob_bytes = blob_file_number << 20;
|
|
|
|
std::string checksum_method(checksum_method_prefix);
|
|
AppendNumberTo(&checksum_method, blob_file_number);
|
|
|
|
std::string checksum_value(checksum_value_prefix);
|
|
AppendNumberTo(&checksum_value, blob_file_number);
|
|
|
|
edit.AddBlobFile(blob_file_number, total_blob_count, total_blob_bytes,
|
|
checksum_method, checksum_value);
|
|
|
|
const uint64_t garbage_blob_count = total_blob_count >> 2;
|
|
const uint64_t garbage_blob_bytes = total_blob_bytes >> 1;
|
|
|
|
edit.AddBlobFileGarbage(blob_file_number, garbage_blob_count,
|
|
garbage_blob_bytes);
|
|
}
|
|
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, AddWalEncodeDecode) {
|
|
VersionEdit edit;
|
|
for (uint64_t log_number = 1; log_number <= 20; log_number++) {
|
|
WalMetadata meta;
|
|
bool has_size = rand() % 2 == 0;
|
|
if (has_size) {
|
|
meta.SetSyncedSizeInBytes(rand() % 1000);
|
|
}
|
|
edit.AddWal(log_number, meta);
|
|
}
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
static std::string PrefixEncodedWalAdditionWithLength(
|
|
const std::string& encoded) {
|
|
std::string ret;
|
|
PutVarint32(&ret, Tag::kWalAddition2);
|
|
PutLengthPrefixedSlice(&ret, encoded);
|
|
return ret;
|
|
}
|
|
|
|
TEST_F(VersionEditTest, AddWalDecodeBadLogNumber) {
|
|
std::string encoded;
|
|
|
|
{
|
|
// No log number.
|
|
std::string encoded_edit = PrefixEncodedWalAdditionWithLength(encoded);
|
|
VersionEdit edit;
|
|
Status s = edit.DecodeFrom(encoded_edit);
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
ASSERT_TRUE(s.ToString().find("Error decoding WAL log number") !=
|
|
std::string::npos)
|
|
<< s.ToString();
|
|
}
|
|
|
|
{
|
|
// log number should be varint64,
|
|
// but we only encode 128 which is not a valid representation of varint64.
|
|
char c = 0;
|
|
unsigned char* ptr = reinterpret_cast<unsigned char*>(&c);
|
|
*ptr = 128;
|
|
encoded.append(1, c);
|
|
|
|
std::string encoded_edit = PrefixEncodedWalAdditionWithLength(encoded);
|
|
VersionEdit edit;
|
|
Status s = edit.DecodeFrom(encoded_edit);
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
ASSERT_TRUE(s.ToString().find("Error decoding WAL log number") !=
|
|
std::string::npos)
|
|
<< s.ToString();
|
|
}
|
|
}
|
|
|
|
TEST_F(VersionEditTest, AddWalDecodeBadTag) {
|
|
constexpr WalNumber kLogNumber = 100;
|
|
constexpr uint64_t kSizeInBytes = 100;
|
|
|
|
std::string encoded;
|
|
PutVarint64(&encoded, kLogNumber);
|
|
|
|
{
|
|
// No tag.
|
|
std::string encoded_edit = PrefixEncodedWalAdditionWithLength(encoded);
|
|
VersionEdit edit;
|
|
Status s = edit.DecodeFrom(encoded_edit);
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
ASSERT_TRUE(s.ToString().find("Error decoding tag") != std::string::npos)
|
|
<< s.ToString();
|
|
}
|
|
|
|
{
|
|
// Only has size tag, no terminate tag.
|
|
std::string encoded_with_size = encoded;
|
|
PutVarint32(&encoded_with_size,
|
|
static_cast<uint32_t>(WalAdditionTag::kSyncedSize));
|
|
PutVarint64(&encoded_with_size, kSizeInBytes);
|
|
|
|
std::string encoded_edit =
|
|
PrefixEncodedWalAdditionWithLength(encoded_with_size);
|
|
VersionEdit edit;
|
|
Status s = edit.DecodeFrom(encoded_edit);
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
ASSERT_TRUE(s.ToString().find("Error decoding tag") != std::string::npos)
|
|
<< s.ToString();
|
|
}
|
|
|
|
{
|
|
// Only has terminate tag.
|
|
std::string encoded_with_terminate = encoded;
|
|
PutVarint32(&encoded_with_terminate,
|
|
static_cast<uint32_t>(WalAdditionTag::kTerminate));
|
|
|
|
std::string encoded_edit =
|
|
PrefixEncodedWalAdditionWithLength(encoded_with_terminate);
|
|
VersionEdit edit;
|
|
ASSERT_OK(edit.DecodeFrom(encoded_edit));
|
|
auto& wal_addition = edit.GetWalAdditions()[0];
|
|
ASSERT_EQ(wal_addition.GetLogNumber(), kLogNumber);
|
|
ASSERT_FALSE(wal_addition.GetMetadata().HasSyncedSize());
|
|
}
|
|
}
|
|
|
|
TEST_F(VersionEditTest, AddWalDecodeNoSize) {
|
|
constexpr WalNumber kLogNumber = 100;
|
|
|
|
std::string encoded;
|
|
PutVarint64(&encoded, kLogNumber);
|
|
PutVarint32(&encoded, static_cast<uint32_t>(WalAdditionTag::kSyncedSize));
|
|
// No real size after the size tag.
|
|
|
|
{
|
|
// Without terminate tag.
|
|
std::string encoded_edit = PrefixEncodedWalAdditionWithLength(encoded);
|
|
VersionEdit edit;
|
|
Status s = edit.DecodeFrom(encoded_edit);
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
ASSERT_TRUE(s.ToString().find("Error decoding WAL file size") !=
|
|
std::string::npos)
|
|
<< s.ToString();
|
|
}
|
|
|
|
{
|
|
// With terminate tag.
|
|
PutVarint32(&encoded, static_cast<uint32_t>(WalAdditionTag::kTerminate));
|
|
|
|
std::string encoded_edit = PrefixEncodedWalAdditionWithLength(encoded);
|
|
VersionEdit edit;
|
|
Status s = edit.DecodeFrom(encoded_edit);
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
// The terminate tag is misunderstood as the size.
|
|
ASSERT_TRUE(s.ToString().find("Error decoding tag") != std::string::npos)
|
|
<< s.ToString();
|
|
}
|
|
}
|
|
|
|
TEST_F(VersionEditTest, AddWalDebug) {
|
|
constexpr int n = 2;
|
|
constexpr std::array<uint64_t, n> kLogNumbers{{10, 20}};
|
|
constexpr std::array<uint64_t, n> kSizeInBytes{{100, 200}};
|
|
|
|
VersionEdit edit;
|
|
for (int i = 0; i < n; i++) {
|
|
edit.AddWal(kLogNumbers[i], WalMetadata(kSizeInBytes[i]));
|
|
}
|
|
|
|
const WalAdditions& wals = edit.GetWalAdditions();
|
|
|
|
ASSERT_TRUE(edit.IsWalAddition());
|
|
ASSERT_EQ(wals.size(), n);
|
|
for (int i = 0; i < n; i++) {
|
|
const WalAddition& wal = wals[i];
|
|
ASSERT_EQ(wal.GetLogNumber(), kLogNumbers[i]);
|
|
ASSERT_EQ(wal.GetMetadata().GetSyncedSizeInBytes(), kSizeInBytes[i]);
|
|
}
|
|
|
|
std::string expected_str = "VersionEdit {\n";
|
|
for (int i = 0; i < n; i++) {
|
|
std::stringstream ss;
|
|
ss << " WalAddition: log_number: " << kLogNumbers[i]
|
|
<< " synced_size_in_bytes: " << kSizeInBytes[i] << "\n";
|
|
expected_str += ss.str();
|
|
}
|
|
expected_str += " ColumnFamily: 0\n}\n";
|
|
ASSERT_EQ(edit.DebugString(true), expected_str);
|
|
|
|
std::string expected_json = "{\"EditNumber\": 4, \"WalAdditions\": [";
|
|
for (int i = 0; i < n; i++) {
|
|
std::stringstream ss;
|
|
ss << "{\"LogNumber\": " << kLogNumbers[i] << ", "
|
|
<< "\"SyncedSizeInBytes\": " << kSizeInBytes[i] << "}";
|
|
if (i < n - 1) ss << ", ";
|
|
expected_json += ss.str();
|
|
}
|
|
expected_json += "], \"ColumnFamily\": 0}";
|
|
ASSERT_EQ(edit.DebugJSON(4, true), expected_json);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, DeleteWalEncodeDecode) {
|
|
VersionEdit edit;
|
|
edit.DeleteWalsBefore(rand() % 100);
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, DeleteWalDebug) {
|
|
constexpr int n = 2;
|
|
constexpr std::array<uint64_t, n> kLogNumbers{{10, 20}};
|
|
|
|
VersionEdit edit;
|
|
edit.DeleteWalsBefore(kLogNumbers[n - 1]);
|
|
|
|
const WalDeletion& wal = edit.GetWalDeletion();
|
|
|
|
ASSERT_TRUE(edit.IsWalDeletion());
|
|
ASSERT_EQ(wal.GetLogNumber(), kLogNumbers[n - 1]);
|
|
|
|
std::string expected_str = "VersionEdit {\n";
|
|
{
|
|
std::stringstream ss;
|
|
ss << " WalDeletion: log_number: " << kLogNumbers[n - 1] << "\n";
|
|
expected_str += ss.str();
|
|
}
|
|
expected_str += " ColumnFamily: 0\n}\n";
|
|
ASSERT_EQ(edit.DebugString(true), expected_str);
|
|
|
|
std::string expected_json = "{\"EditNumber\": 4, \"WalDeletion\": ";
|
|
{
|
|
std::stringstream ss;
|
|
ss << "{\"LogNumber\": " << kLogNumbers[n - 1] << "}";
|
|
expected_json += ss.str();
|
|
}
|
|
expected_json += ", \"ColumnFamily\": 0}";
|
|
ASSERT_EQ(edit.DebugJSON(4, true), expected_json);
|
|
}
|
|
|
|
TEST_F(VersionEditTest, FullHistoryTsLow) {
|
|
VersionEdit edit;
|
|
ASSERT_FALSE(edit.HasFullHistoryTsLow());
|
|
std::string ts = test::EncodeInt(0);
|
|
edit.SetFullHistoryTsLow(ts);
|
|
TestEncodeDecode(edit);
|
|
}
|
|
|
|
// Tests that if RocksDB is downgraded, the new types of VersionEdits
|
|
// that have a tag larger than kTagSafeIgnoreMask can be safely ignored.
|
|
TEST_F(VersionEditTest, IgnorableTags) {
|
|
SyncPoint::GetInstance()->SetCallBack(
|
|
"VersionEdit::EncodeTo:IgnoreIgnorableTags", [&](void* arg) {
|
|
bool* ignore = static_cast<bool*>(arg);
|
|
*ignore = true;
|
|
});
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
constexpr uint64_t kPrevLogNumber = 100;
|
|
constexpr uint64_t kLogNumber = 200;
|
|
constexpr uint64_t kNextFileNumber = 300;
|
|
constexpr uint64_t kColumnFamilyId = 400;
|
|
|
|
VersionEdit edit;
|
|
// Add some ignorable entries.
|
|
for (int i = 0; i < 2; i++) {
|
|
edit.AddWal(i + 1, WalMetadata(i + 2));
|
|
}
|
|
edit.SetDBId("db_id");
|
|
// Add unignorable entries.
|
|
edit.SetPrevLogNumber(kPrevLogNumber);
|
|
edit.SetLogNumber(kLogNumber);
|
|
// Add more ignorable entries.
|
|
edit.DeleteWalsBefore(100);
|
|
// Add unignorable entry.
|
|
edit.SetNextFile(kNextFileNumber);
|
|
// Add more ignorable entries.
|
|
edit.SetFullHistoryTsLow("ts");
|
|
// Add unignorable entry.
|
|
edit.SetColumnFamily(kColumnFamilyId);
|
|
|
|
std::string encoded;
|
|
ASSERT_TRUE(edit.EncodeTo(&encoded));
|
|
|
|
VersionEdit decoded;
|
|
ASSERT_OK(decoded.DecodeFrom(encoded));
|
|
|
|
// Check that all ignorable entries are ignored.
|
|
ASSERT_FALSE(decoded.HasDbId());
|
|
ASSERT_FALSE(decoded.HasFullHistoryTsLow());
|
|
ASSERT_FALSE(decoded.IsWalAddition());
|
|
ASSERT_FALSE(decoded.IsWalDeletion());
|
|
ASSERT_TRUE(decoded.GetWalAdditions().empty());
|
|
ASSERT_TRUE(decoded.GetWalDeletion().IsEmpty());
|
|
|
|
// Check that unignorable entries are still present.
|
|
ASSERT_EQ(edit.GetPrevLogNumber(), kPrevLogNumber);
|
|
ASSERT_EQ(edit.GetLogNumber(), kLogNumber);
|
|
ASSERT_EQ(edit.GetNextFile(), kNextFileNumber);
|
|
ASSERT_EQ(edit.GetColumnFamily(), kColumnFamilyId);
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
}
|
|
|
|
TEST(FileMetaDataTest, UpdateBoundariesBlobIndex) {
|
|
FileMetaData meta;
|
|
|
|
{
|
|
constexpr uint64_t file_number = 10;
|
|
constexpr uint32_t path_id = 0;
|
|
constexpr uint64_t file_size = 0;
|
|
|
|
meta.fd = FileDescriptor(file_number, path_id, file_size);
|
|
}
|
|
|
|
constexpr char key[] = "foo";
|
|
|
|
constexpr uint64_t expected_oldest_blob_file_number = 20;
|
|
|
|
// Plain old value (does not affect oldest_blob_file_number)
|
|
{
|
|
constexpr char value[] = "value";
|
|
constexpr SequenceNumber seq = 200;
|
|
|
|
ASSERT_OK(meta.UpdateBoundaries(key, value, seq, kTypeValue));
|
|
ASSERT_EQ(meta.oldest_blob_file_number, kInvalidBlobFileNumber);
|
|
}
|
|
|
|
// Non-inlined, non-TTL blob index (sets oldest_blob_file_number)
|
|
{
|
|
constexpr uint64_t blob_file_number = 25;
|
|
static_assert(blob_file_number > expected_oldest_blob_file_number,
|
|
"unexpected");
|
|
|
|
constexpr uint64_t offset = 1000;
|
|
constexpr uint64_t size = 100;
|
|
|
|
std::string blob_index;
|
|
BlobIndex::EncodeBlob(&blob_index, blob_file_number, offset, size,
|
|
kNoCompression);
|
|
|
|
constexpr SequenceNumber seq = 201;
|
|
|
|
ASSERT_OK(meta.UpdateBoundaries(key, blob_index, seq, kTypeBlobIndex));
|
|
ASSERT_EQ(meta.oldest_blob_file_number, blob_file_number);
|
|
}
|
|
|
|
// Another one, with the oldest blob file number (updates
|
|
// oldest_blob_file_number)
|
|
{
|
|
constexpr uint64_t offset = 2000;
|
|
constexpr uint64_t size = 300;
|
|
|
|
std::string blob_index;
|
|
BlobIndex::EncodeBlob(&blob_index, expected_oldest_blob_file_number, offset,
|
|
size, kNoCompression);
|
|
|
|
constexpr SequenceNumber seq = 202;
|
|
|
|
ASSERT_OK(meta.UpdateBoundaries(key, blob_index, seq, kTypeBlobIndex));
|
|
ASSERT_EQ(meta.oldest_blob_file_number, expected_oldest_blob_file_number);
|
|
}
|
|
|
|
// Inlined TTL blob index (does not affect oldest_blob_file_number)
|
|
{
|
|
constexpr uint64_t expiration = 9876543210;
|
|
constexpr char value[] = "value";
|
|
|
|
std::string blob_index;
|
|
BlobIndex::EncodeInlinedTTL(&blob_index, expiration, value);
|
|
|
|
constexpr SequenceNumber seq = 203;
|
|
|
|
ASSERT_OK(meta.UpdateBoundaries(key, blob_index, seq, kTypeBlobIndex));
|
|
ASSERT_EQ(meta.oldest_blob_file_number, expected_oldest_blob_file_number);
|
|
}
|
|
|
|
// Non-inlined TTL blob index (does not affect oldest_blob_file_number, even
|
|
// though file number is smaller)
|
|
{
|
|
constexpr uint64_t expiration = 9876543210;
|
|
constexpr uint64_t blob_file_number = 15;
|
|
static_assert(blob_file_number < expected_oldest_blob_file_number,
|
|
"unexpected");
|
|
|
|
constexpr uint64_t offset = 2000;
|
|
constexpr uint64_t size = 500;
|
|
|
|
std::string blob_index;
|
|
BlobIndex::EncodeBlobTTL(&blob_index, expiration, blob_file_number, offset,
|
|
size, kNoCompression);
|
|
|
|
constexpr SequenceNumber seq = 204;
|
|
|
|
ASSERT_OK(meta.UpdateBoundaries(key, blob_index, seq, kTypeBlobIndex));
|
|
ASSERT_EQ(meta.oldest_blob_file_number, expected_oldest_blob_file_number);
|
|
}
|
|
|
|
// Corrupt blob index
|
|
{
|
|
constexpr char corrupt_blob_index[] = "!corrupt!";
|
|
constexpr SequenceNumber seq = 205;
|
|
|
|
ASSERT_TRUE(
|
|
meta.UpdateBoundaries(key, corrupt_blob_index, seq, kTypeBlobIndex)
|
|
.IsCorruption());
|
|
ASSERT_EQ(meta.oldest_blob_file_number, expected_oldest_blob_file_number);
|
|
}
|
|
|
|
// Invalid blob file number
|
|
{
|
|
constexpr uint64_t offset = 10000;
|
|
constexpr uint64_t size = 1000;
|
|
|
|
std::string blob_index;
|
|
BlobIndex::EncodeBlob(&blob_index, kInvalidBlobFileNumber, offset, size,
|
|
kNoCompression);
|
|
|
|
constexpr SequenceNumber seq = 206;
|
|
|
|
ASSERT_TRUE(meta.UpdateBoundaries(key, blob_index, seq, kTypeBlobIndex)
|
|
.IsCorruption());
|
|
ASSERT_EQ(meta.oldest_blob_file_number, expected_oldest_blob_file_number);
|
|
}
|
|
}
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
int main(int argc, char** argv) {
|
|
ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
return RUN_ALL_TESTS();
|
|
}
|