mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-26 07:30:54 +00:00
6a0f637633
Summary: ... to improve data integrity validation during compaction. A new option `compaction_verify_record_count` is introduced for this verification and is enabled by default. One exception when the verification is not done is when a compaction filter returns kRemoveAndSkipUntil which can cause CompactionIterator to seek until some key and hence not able to keep track of the number of keys processed. For expected number of input keys, we sum over the number of total keys - number of range tombstones across compaction input files (`CompactionJob::UpdateCompactionStats()`). Table properties are consulted if `FileMetaData` is not initialized for some input file. Since table properties for all input files were also constructed during `DBImpl::NotifyOnCompactionBegin()`, `Compaction::GetTableProperties()` is introduced to reduce duplicated code. For actual number of keys processed, each subcompaction will record its number of keys processed to `sub_compact->compaction_job_stats.num_input_records` and aggregated when all subcompactions finish (`CompactionJob::AggregateCompactionStats()`). In the case when some subcompaction encountered kRemoveAndSkipUntil from compaction filter and does not have accurate count, it propagates this information through `sub_compact->compaction_job_stats.has_num_input_records`. Pull Request resolved: https://github.com/facebook/rocksdb/pull/11571 Test Plan: * Add a new unit test `DBCompactionTest.VerifyRecordCount` for the corruption case. * All other unit tests for non-corrupted case. * Ran crash test for a few hours: `python3 ./tools/db_crashtest.py whitebox --simple` Reviewed By: ajkr Differential Revision: D47131965 Pulled By: cbi42 fbshipit-source-id: cc8e94565dd526c4347e9d3843ecf32f6727af92
745 lines
28 KiB
C++
745 lines
28 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#pragma once
|
|
#include <algorithm>
|
|
#include <optional>
|
|
#include <set>
|
|
#include <string>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include "db/blob/blob_file_addition.h"
|
|
#include "db/blob/blob_file_garbage.h"
|
|
#include "db/dbformat.h"
|
|
#include "db/wal_edit.h"
|
|
#include "memory/arena.h"
|
|
#include "port/malloc.h"
|
|
#include "rocksdb/advanced_cache.h"
|
|
#include "rocksdb/advanced_options.h"
|
|
#include "table/table_reader.h"
|
|
#include "table/unique_id_impl.h"
|
|
#include "util/autovector.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
// Tag numbers for serialized VersionEdit. These numbers are written to
|
|
// disk and should not be changed. The number should be forward compatible so
|
|
// users can down-grade RocksDB safely. A future Tag is ignored by doing '&'
|
|
// between Tag and kTagSafeIgnoreMask field.
|
|
enum Tag : uint32_t {
|
|
kComparator = 1,
|
|
kLogNumber = 2,
|
|
kNextFileNumber = 3,
|
|
kLastSequence = 4,
|
|
kCompactCursor = 5,
|
|
kDeletedFile = 6,
|
|
kNewFile = 7,
|
|
// 8 was used for large value refs
|
|
kPrevLogNumber = 9,
|
|
kMinLogNumberToKeep = 10,
|
|
|
|
// these are new formats divergent from open source leveldb
|
|
kNewFile2 = 100,
|
|
kNewFile3 = 102,
|
|
kNewFile4 = 103, // 4th (the latest) format version of adding files
|
|
kColumnFamily = 200, // specify column family for version edit
|
|
kColumnFamilyAdd = 201,
|
|
kColumnFamilyDrop = 202,
|
|
kMaxColumnFamily = 203,
|
|
|
|
kInAtomicGroup = 300,
|
|
|
|
kBlobFileAddition = 400,
|
|
kBlobFileGarbage,
|
|
|
|
// Mask for an unidentified tag from the future which can be safely ignored.
|
|
kTagSafeIgnoreMask = 1 << 13,
|
|
|
|
// Forward compatible (aka ignorable) records
|
|
kDbId,
|
|
kBlobFileAddition_DEPRECATED,
|
|
kBlobFileGarbage_DEPRECATED,
|
|
kWalAddition,
|
|
kWalDeletion,
|
|
kFullHistoryTsLow,
|
|
kWalAddition2,
|
|
kWalDeletion2,
|
|
kPersistUserDefinedTimestamps,
|
|
};
|
|
|
|
enum NewFileCustomTag : uint32_t {
|
|
kTerminate = 1, // The end of customized fields
|
|
kNeedCompaction = 2,
|
|
// Since Manifest is not entirely forward-compatible, we currently encode
|
|
// kMinLogNumberToKeep as part of NewFile as a hack. This should be removed
|
|
// when manifest becomes forward-compatible.
|
|
kMinLogNumberToKeepHack = 3,
|
|
kOldestBlobFileNumber = 4,
|
|
kOldestAncesterTime = 5,
|
|
kFileCreationTime = 6,
|
|
kFileChecksum = 7,
|
|
kFileChecksumFuncName = 8,
|
|
kTemperature = 9,
|
|
kMinTimestamp = 10,
|
|
kMaxTimestamp = 11,
|
|
kUniqueId = 12,
|
|
kEpochNumber = 13,
|
|
kCompensatedRangeDeletionSize = 14,
|
|
kTailSize = 15,
|
|
kUserDefinedTimestampsPersisted = 16,
|
|
|
|
// If this bit for the custom tag is set, opening DB should fail if
|
|
// we don't know this field.
|
|
kCustomTagNonSafeIgnoreMask = 1 << 6,
|
|
|
|
// Forward incompatible (aka unignorable) fields
|
|
kPathId,
|
|
};
|
|
|
|
class VersionSet;
|
|
|
|
constexpr uint64_t kFileNumberMask = 0x3FFFFFFFFFFFFFFF;
|
|
constexpr uint64_t kUnknownOldestAncesterTime = 0;
|
|
constexpr uint64_t kUnknownFileCreationTime = 0;
|
|
constexpr uint64_t kUnknownEpochNumber = 0;
|
|
// If `Options::allow_ingest_behind` is true, this epoch number
|
|
// will be dedicated to files ingested behind.
|
|
constexpr uint64_t kReservedEpochNumberForFileIngestedBehind = 1;
|
|
|
|
extern uint64_t PackFileNumberAndPathId(uint64_t number, uint64_t path_id);
|
|
|
|
// A copyable structure contains information needed to read data from an SST
|
|
// file. It can contain a pointer to a table reader opened for the file, or
|
|
// file number and size, which can be used to create a new table reader for it.
|
|
// The behavior is undefined when a copied of the structure is used when the
|
|
// file is not in any live version any more.
|
|
struct FileDescriptor {
|
|
// Table reader in table_reader_handle
|
|
TableReader* table_reader;
|
|
uint64_t packed_number_and_path_id;
|
|
uint64_t file_size; // File size in bytes
|
|
SequenceNumber smallest_seqno; // The smallest seqno in this file
|
|
SequenceNumber largest_seqno; // The largest seqno in this file
|
|
|
|
FileDescriptor() : FileDescriptor(0, 0, 0) {}
|
|
|
|
FileDescriptor(uint64_t number, uint32_t path_id, uint64_t _file_size)
|
|
: FileDescriptor(number, path_id, _file_size, kMaxSequenceNumber, 0) {}
|
|
|
|
FileDescriptor(uint64_t number, uint32_t path_id, uint64_t _file_size,
|
|
SequenceNumber _smallest_seqno, SequenceNumber _largest_seqno)
|
|
: table_reader(nullptr),
|
|
packed_number_and_path_id(PackFileNumberAndPathId(number, path_id)),
|
|
file_size(_file_size),
|
|
smallest_seqno(_smallest_seqno),
|
|
largest_seqno(_largest_seqno) {}
|
|
|
|
FileDescriptor(const FileDescriptor& fd) { *this = fd; }
|
|
|
|
FileDescriptor& operator=(const FileDescriptor& fd) {
|
|
table_reader = fd.table_reader;
|
|
packed_number_and_path_id = fd.packed_number_and_path_id;
|
|
file_size = fd.file_size;
|
|
smallest_seqno = fd.smallest_seqno;
|
|
largest_seqno = fd.largest_seqno;
|
|
return *this;
|
|
}
|
|
|
|
uint64_t GetNumber() const {
|
|
return packed_number_and_path_id & kFileNumberMask;
|
|
}
|
|
uint32_t GetPathId() const {
|
|
return static_cast<uint32_t>(packed_number_and_path_id /
|
|
(kFileNumberMask + 1));
|
|
}
|
|
uint64_t GetFileSize() const { return file_size; }
|
|
};
|
|
|
|
struct FileSampledStats {
|
|
FileSampledStats() : num_reads_sampled(0) {}
|
|
FileSampledStats(const FileSampledStats& other) { *this = other; }
|
|
FileSampledStats& operator=(const FileSampledStats& other) {
|
|
num_reads_sampled = other.num_reads_sampled.load();
|
|
return *this;
|
|
}
|
|
|
|
// number of user reads to this file.
|
|
mutable std::atomic<uint64_t> num_reads_sampled;
|
|
};
|
|
|
|
struct FileMetaData {
|
|
FileDescriptor fd;
|
|
InternalKey smallest; // Smallest internal key served by table
|
|
InternalKey largest; // Largest internal key served by table
|
|
|
|
// Needs to be disposed when refs becomes 0.
|
|
Cache::Handle* table_reader_handle = nullptr;
|
|
|
|
FileSampledStats stats;
|
|
|
|
// Stats for compensating deletion entries during compaction
|
|
|
|
// File size compensated by deletion entry.
|
|
// This is used to compute a file's compaction priority, and is updated in
|
|
// Version::ComputeCompensatedSizes() first time when the file is created or
|
|
// loaded. After it is updated (!= 0), it is immutable.
|
|
uint64_t compensated_file_size = 0;
|
|
// These values can mutate, but they can only be read or written from
|
|
// single-threaded LogAndApply thread
|
|
uint64_t num_entries =
|
|
0; // The number of entries, including deletions and range deletions.
|
|
// The number of deletion entries, including range deletions.
|
|
uint64_t num_deletions = 0;
|
|
uint64_t raw_key_size = 0; // total uncompressed key size.
|
|
uint64_t raw_value_size = 0; // total uncompressed value size.
|
|
uint64_t num_range_deletions = 0;
|
|
// This is computed during Flush/Compaction, and is added to
|
|
// `compensated_file_size`. Currently, this estimates the size of keys in the
|
|
// next level covered by range tombstones in this file.
|
|
uint64_t compensated_range_deletion_size = 0;
|
|
|
|
int refs = 0; // Reference count
|
|
|
|
bool being_compacted = false; // Is this file undergoing compaction?
|
|
bool init_stats_from_file = false; // true if the data-entry stats of this
|
|
// file has initialized from file.
|
|
|
|
bool marked_for_compaction = false; // True if client asked us nicely to
|
|
// compact this file.
|
|
Temperature temperature = Temperature::kUnknown;
|
|
|
|
// Used only in BlobDB. The file number of the oldest blob file this SST file
|
|
// refers to. 0 is an invalid value; BlobDB numbers the files starting from 1.
|
|
uint64_t oldest_blob_file_number = kInvalidBlobFileNumber;
|
|
|
|
// The file could be the compaction output from other SST files, which could
|
|
// in turn be outputs for compact older SST files. We track the memtable
|
|
// flush timestamp for the oldest SST file that eventually contribute data
|
|
// to this file. 0 means the information is not available.
|
|
uint64_t oldest_ancester_time = kUnknownOldestAncesterTime;
|
|
|
|
// Unix time when the SST file is created.
|
|
uint64_t file_creation_time = kUnknownFileCreationTime;
|
|
|
|
// The order of a file being flushed or ingested/imported.
|
|
// Compaction output file will be assigned with the minimum `epoch_number`
|
|
// among input files'.
|
|
// For L0, larger `epoch_number` indicates newer L0 file.
|
|
uint64_t epoch_number = kUnknownEpochNumber;
|
|
|
|
// File checksum
|
|
std::string file_checksum = kUnknownFileChecksum;
|
|
|
|
// File checksum function name
|
|
std::string file_checksum_func_name = kUnknownFileChecksumFuncName;
|
|
|
|
// SST unique id
|
|
UniqueId64x2 unique_id{};
|
|
|
|
// Size of the "tail" part of a SST file
|
|
// "Tail" refers to all blocks after data blocks till the end of the SST file
|
|
uint64_t tail_size = 0;
|
|
|
|
// Value of the `AdvancedColumnFamilyOptions.persist_user_defined_timestamps`
|
|
// flag when the file is created. Default to true, only when this flag is
|
|
// false, it's explicitly written to Manifest.
|
|
bool user_defined_timestamps_persisted = true;
|
|
|
|
FileMetaData() = default;
|
|
|
|
FileMetaData(uint64_t file, uint32_t file_path_id, uint64_t file_size,
|
|
const InternalKey& smallest_key, const InternalKey& largest_key,
|
|
const SequenceNumber& smallest_seq,
|
|
const SequenceNumber& largest_seq, bool marked_for_compact,
|
|
Temperature _temperature, uint64_t oldest_blob_file,
|
|
uint64_t _oldest_ancester_time, uint64_t _file_creation_time,
|
|
uint64_t _epoch_number, const std::string& _file_checksum,
|
|
const std::string& _file_checksum_func_name,
|
|
UniqueId64x2 _unique_id,
|
|
const uint64_t _compensated_range_deletion_size,
|
|
uint64_t _tail_size, bool _user_defined_timestamps_persisted)
|
|
: fd(file, file_path_id, file_size, smallest_seq, largest_seq),
|
|
smallest(smallest_key),
|
|
largest(largest_key),
|
|
compensated_range_deletion_size(_compensated_range_deletion_size),
|
|
marked_for_compaction(marked_for_compact),
|
|
temperature(_temperature),
|
|
oldest_blob_file_number(oldest_blob_file),
|
|
oldest_ancester_time(_oldest_ancester_time),
|
|
file_creation_time(_file_creation_time),
|
|
epoch_number(_epoch_number),
|
|
file_checksum(_file_checksum),
|
|
file_checksum_func_name(_file_checksum_func_name),
|
|
unique_id(std::move(_unique_id)),
|
|
tail_size(_tail_size),
|
|
user_defined_timestamps_persisted(_user_defined_timestamps_persisted) {
|
|
TEST_SYNC_POINT_CALLBACK("FileMetaData::FileMetaData", this);
|
|
}
|
|
|
|
// REQUIRED: Keys must be given to the function in sorted order (it expects
|
|
// the last key to be the largest).
|
|
Status UpdateBoundaries(const Slice& key, const Slice& value,
|
|
SequenceNumber seqno, ValueType value_type);
|
|
|
|
// Unlike UpdateBoundaries, ranges do not need to be presented in any
|
|
// particular order.
|
|
void UpdateBoundariesForRange(const InternalKey& start,
|
|
const InternalKey& end, SequenceNumber seqno,
|
|
const InternalKeyComparator& icmp) {
|
|
if (smallest.size() == 0 || icmp.Compare(start, smallest) < 0) {
|
|
smallest = start;
|
|
}
|
|
if (largest.size() == 0 || icmp.Compare(largest, end) < 0) {
|
|
largest = end;
|
|
}
|
|
assert(icmp.Compare(smallest, largest) <= 0);
|
|
fd.smallest_seqno = std::min(fd.smallest_seqno, seqno);
|
|
fd.largest_seqno = std::max(fd.largest_seqno, seqno);
|
|
}
|
|
|
|
// Try to get oldest ancester time from the class itself or table properties
|
|
// if table reader is already pinned.
|
|
// 0 means the information is not available.
|
|
uint64_t TryGetOldestAncesterTime() {
|
|
if (oldest_ancester_time != kUnknownOldestAncesterTime) {
|
|
return oldest_ancester_time;
|
|
} else if (fd.table_reader != nullptr &&
|
|
fd.table_reader->GetTableProperties() != nullptr) {
|
|
return fd.table_reader->GetTableProperties()->creation_time;
|
|
}
|
|
return kUnknownOldestAncesterTime;
|
|
}
|
|
|
|
uint64_t TryGetFileCreationTime() {
|
|
if (file_creation_time != kUnknownFileCreationTime) {
|
|
return file_creation_time;
|
|
} else if (fd.table_reader != nullptr &&
|
|
fd.table_reader->GetTableProperties() != nullptr) {
|
|
return fd.table_reader->GetTableProperties()->file_creation_time;
|
|
}
|
|
return kUnknownFileCreationTime;
|
|
}
|
|
|
|
// WARNING: manual update to this function is needed
|
|
// whenever a new string property is added to FileMetaData
|
|
// to reduce approximation error.
|
|
//
|
|
// TODO: eliminate the need of manually updating this function
|
|
// for new string properties
|
|
size_t ApproximateMemoryUsage() const {
|
|
size_t usage = 0;
|
|
#ifdef ROCKSDB_MALLOC_USABLE_SIZE
|
|
usage += malloc_usable_size(const_cast<FileMetaData*>(this));
|
|
#else
|
|
usage += sizeof(*this);
|
|
#endif // ROCKSDB_MALLOC_USABLE_SIZE
|
|
usage += smallest.size() + largest.size() + file_checksum.size() +
|
|
file_checksum_func_name.size();
|
|
return usage;
|
|
}
|
|
};
|
|
|
|
// A compressed copy of file meta data that just contain minimum data needed
|
|
// to serve read operations, while still keeping the pointer to full metadata
|
|
// of the file in case it is needed.
|
|
struct FdWithKeyRange {
|
|
FileDescriptor fd;
|
|
FileMetaData* file_metadata; // Point to all metadata
|
|
Slice smallest_key; // slice that contain smallest key
|
|
Slice largest_key; // slice that contain largest key
|
|
|
|
FdWithKeyRange()
|
|
: fd(), file_metadata(nullptr), smallest_key(), largest_key() {}
|
|
|
|
FdWithKeyRange(FileDescriptor _fd, Slice _smallest_key, Slice _largest_key,
|
|
FileMetaData* _file_metadata)
|
|
: fd(_fd),
|
|
file_metadata(_file_metadata),
|
|
smallest_key(_smallest_key),
|
|
largest_key(_largest_key) {}
|
|
};
|
|
|
|
// Data structure to store an array of FdWithKeyRange in one level
|
|
// Actual data is guaranteed to be stored closely
|
|
struct LevelFilesBrief {
|
|
size_t num_files;
|
|
FdWithKeyRange* files;
|
|
LevelFilesBrief() {
|
|
num_files = 0;
|
|
files = nullptr;
|
|
}
|
|
};
|
|
|
|
// The state of a DB at any given time is referred to as a Version.
|
|
// Any modification to the Version is considered a Version Edit. A Version is
|
|
// constructed by joining a sequence of Version Edits. Version Edits are written
|
|
// to the MANIFEST file.
|
|
class VersionEdit {
|
|
public:
|
|
void Clear();
|
|
|
|
void SetDBId(const std::string& db_id) {
|
|
has_db_id_ = true;
|
|
db_id_ = db_id;
|
|
}
|
|
bool HasDbId() const { return has_db_id_; }
|
|
const std::string& GetDbId() const { return db_id_; }
|
|
|
|
void SetComparatorName(const Slice& name) {
|
|
has_comparator_ = true;
|
|
comparator_ = name.ToString();
|
|
}
|
|
bool HasComparatorName() const { return has_comparator_; }
|
|
const std::string& GetComparatorName() const { return comparator_; }
|
|
|
|
void SetPersistUserDefinedTimestamps(bool persist_user_defined_timestamps) {
|
|
has_persist_user_defined_timestamps_ = true;
|
|
persist_user_defined_timestamps_ = persist_user_defined_timestamps;
|
|
}
|
|
bool HasPersistUserDefinedTimestamps() const {
|
|
return has_persist_user_defined_timestamps_;
|
|
}
|
|
bool GetPersistUserDefinedTimestamps() const {
|
|
return persist_user_defined_timestamps_;
|
|
}
|
|
|
|
void SetLogNumber(uint64_t num) {
|
|
has_log_number_ = true;
|
|
log_number_ = num;
|
|
}
|
|
bool HasLogNumber() const { return has_log_number_; }
|
|
uint64_t GetLogNumber() const { return log_number_; }
|
|
|
|
void SetPrevLogNumber(uint64_t num) {
|
|
has_prev_log_number_ = true;
|
|
prev_log_number_ = num;
|
|
}
|
|
bool HasPrevLogNumber() const { return has_prev_log_number_; }
|
|
uint64_t GetPrevLogNumber() const { return prev_log_number_; }
|
|
|
|
void SetNextFile(uint64_t num) {
|
|
has_next_file_number_ = true;
|
|
next_file_number_ = num;
|
|
}
|
|
bool HasNextFile() const { return has_next_file_number_; }
|
|
uint64_t GetNextFile() const { return next_file_number_; }
|
|
|
|
void SetMaxColumnFamily(uint32_t max_column_family) {
|
|
has_max_column_family_ = true;
|
|
max_column_family_ = max_column_family;
|
|
}
|
|
bool HasMaxColumnFamily() const { return has_max_column_family_; }
|
|
uint32_t GetMaxColumnFamily() const { return max_column_family_; }
|
|
|
|
void SetMinLogNumberToKeep(uint64_t num) {
|
|
has_min_log_number_to_keep_ = true;
|
|
min_log_number_to_keep_ = num;
|
|
}
|
|
bool HasMinLogNumberToKeep() const { return has_min_log_number_to_keep_; }
|
|
uint64_t GetMinLogNumberToKeep() const { return min_log_number_to_keep_; }
|
|
|
|
void SetLastSequence(SequenceNumber seq) {
|
|
has_last_sequence_ = true;
|
|
last_sequence_ = seq;
|
|
}
|
|
bool HasLastSequence() const { return has_last_sequence_; }
|
|
SequenceNumber GetLastSequence() const { return last_sequence_; }
|
|
|
|
// Delete the specified table file from the specified level.
|
|
void DeleteFile(int level, uint64_t file) {
|
|
deleted_files_.emplace(level, file);
|
|
}
|
|
|
|
// Retrieve the table files deleted as well as their associated levels.
|
|
using DeletedFiles = std::set<std::pair<int, uint64_t>>;
|
|
const DeletedFiles& GetDeletedFiles() const { return deleted_files_; }
|
|
|
|
// Add the specified table file at the specified level.
|
|
// REQUIRES: "smallest" and "largest" are smallest and largest keys in file
|
|
// REQUIRES: "oldest_blob_file_number" is the number of the oldest blob file
|
|
// referred to by this file if any, kInvalidBlobFileNumber otherwise.
|
|
void AddFile(int level, uint64_t file, uint32_t file_path_id,
|
|
uint64_t file_size, const InternalKey& smallest,
|
|
const InternalKey& largest, const SequenceNumber& smallest_seqno,
|
|
const SequenceNumber& largest_seqno, bool marked_for_compaction,
|
|
Temperature temperature, uint64_t oldest_blob_file_number,
|
|
uint64_t oldest_ancester_time, uint64_t file_creation_time,
|
|
uint64_t epoch_number, const std::string& file_checksum,
|
|
const std::string& file_checksum_func_name,
|
|
const UniqueId64x2& unique_id,
|
|
const uint64_t compensated_range_deletion_size,
|
|
uint64_t tail_size, bool user_defined_timestamps_persisted) {
|
|
assert(smallest_seqno <= largest_seqno);
|
|
new_files_.emplace_back(
|
|
level,
|
|
FileMetaData(file, file_path_id, file_size, smallest, largest,
|
|
smallest_seqno, largest_seqno, marked_for_compaction,
|
|
temperature, oldest_blob_file_number, oldest_ancester_time,
|
|
file_creation_time, epoch_number, file_checksum,
|
|
file_checksum_func_name, unique_id,
|
|
compensated_range_deletion_size, tail_size,
|
|
user_defined_timestamps_persisted));
|
|
if (!HasLastSequence() || largest_seqno > GetLastSequence()) {
|
|
SetLastSequence(largest_seqno);
|
|
}
|
|
}
|
|
|
|
void AddFile(int level, const FileMetaData& f) {
|
|
assert(f.fd.smallest_seqno <= f.fd.largest_seqno);
|
|
new_files_.emplace_back(level, f);
|
|
if (!HasLastSequence() || f.fd.largest_seqno > GetLastSequence()) {
|
|
SetLastSequence(f.fd.largest_seqno);
|
|
}
|
|
}
|
|
|
|
// Retrieve the table files added as well as their associated levels.
|
|
using NewFiles = std::vector<std::pair<int, FileMetaData>>;
|
|
const NewFiles& GetNewFiles() const { return new_files_; }
|
|
|
|
NewFiles& GetMutableNewFiles() { return new_files_; }
|
|
|
|
// Retrieve all the compact cursors
|
|
using CompactCursors = std::vector<std::pair<int, InternalKey>>;
|
|
const CompactCursors& GetCompactCursors() const { return compact_cursors_; }
|
|
void AddCompactCursor(int level, const InternalKey& cursor) {
|
|
compact_cursors_.push_back(std::make_pair(level, cursor));
|
|
}
|
|
void SetCompactCursors(
|
|
const std::vector<InternalKey>& compact_cursors_by_level) {
|
|
compact_cursors_.clear();
|
|
compact_cursors_.reserve(compact_cursors_by_level.size());
|
|
for (int i = 0; i < (int)compact_cursors_by_level.size(); i++) {
|
|
if (compact_cursors_by_level[i].Valid()) {
|
|
compact_cursors_.push_back(
|
|
std::make_pair(i, compact_cursors_by_level[i]));
|
|
}
|
|
}
|
|
}
|
|
|
|
// Add a new blob file.
|
|
void AddBlobFile(uint64_t blob_file_number, uint64_t total_blob_count,
|
|
uint64_t total_blob_bytes, std::string checksum_method,
|
|
std::string checksum_value) {
|
|
blob_file_additions_.emplace_back(
|
|
blob_file_number, total_blob_count, total_blob_bytes,
|
|
std::move(checksum_method), std::move(checksum_value));
|
|
}
|
|
|
|
void AddBlobFile(BlobFileAddition blob_file_addition) {
|
|
blob_file_additions_.emplace_back(std::move(blob_file_addition));
|
|
}
|
|
|
|
// Retrieve all the blob files added.
|
|
using BlobFileAdditions = std::vector<BlobFileAddition>;
|
|
const BlobFileAdditions& GetBlobFileAdditions() const {
|
|
return blob_file_additions_;
|
|
}
|
|
|
|
void SetBlobFileAdditions(BlobFileAdditions blob_file_additions) {
|
|
assert(blob_file_additions_.empty());
|
|
blob_file_additions_ = std::move(blob_file_additions);
|
|
}
|
|
|
|
// Add garbage for an existing blob file. Note: intentionally broken English
|
|
// follows.
|
|
void AddBlobFileGarbage(uint64_t blob_file_number,
|
|
uint64_t garbage_blob_count,
|
|
uint64_t garbage_blob_bytes) {
|
|
blob_file_garbages_.emplace_back(blob_file_number, garbage_blob_count,
|
|
garbage_blob_bytes);
|
|
}
|
|
|
|
void AddBlobFileGarbage(BlobFileGarbage blob_file_garbage) {
|
|
blob_file_garbages_.emplace_back(std::move(blob_file_garbage));
|
|
}
|
|
|
|
// Retrieve all the blob file garbage added.
|
|
using BlobFileGarbages = std::vector<BlobFileGarbage>;
|
|
const BlobFileGarbages& GetBlobFileGarbages() const {
|
|
return blob_file_garbages_;
|
|
}
|
|
|
|
void SetBlobFileGarbages(BlobFileGarbages blob_file_garbages) {
|
|
assert(blob_file_garbages_.empty());
|
|
blob_file_garbages_ = std::move(blob_file_garbages);
|
|
}
|
|
|
|
// Add a WAL (either just created or closed).
|
|
// AddWal and DeleteWalsBefore cannot be called on the same VersionEdit.
|
|
void AddWal(WalNumber number, WalMetadata metadata = WalMetadata()) {
|
|
assert(NumEntries() == wal_additions_.size());
|
|
wal_additions_.emplace_back(number, std::move(metadata));
|
|
}
|
|
|
|
// Retrieve all the added WALs.
|
|
const WalAdditions& GetWalAdditions() const { return wal_additions_; }
|
|
|
|
bool IsWalAddition() const { return !wal_additions_.empty(); }
|
|
|
|
// Delete a WAL (either directly deleted or archived).
|
|
// AddWal and DeleteWalsBefore cannot be called on the same VersionEdit.
|
|
void DeleteWalsBefore(WalNumber number) {
|
|
assert((NumEntries() == 1) == !wal_deletion_.IsEmpty());
|
|
wal_deletion_ = WalDeletion(number);
|
|
}
|
|
|
|
const WalDeletion& GetWalDeletion() const { return wal_deletion_; }
|
|
|
|
bool IsWalDeletion() const { return !wal_deletion_.IsEmpty(); }
|
|
|
|
bool IsWalManipulation() const {
|
|
size_t entries = NumEntries();
|
|
return (entries > 0) && ((entries == wal_additions_.size()) ||
|
|
(entries == !wal_deletion_.IsEmpty()));
|
|
}
|
|
|
|
// Number of edits
|
|
size_t NumEntries() const {
|
|
return new_files_.size() + deleted_files_.size() +
|
|
blob_file_additions_.size() + blob_file_garbages_.size() +
|
|
wal_additions_.size() + !wal_deletion_.IsEmpty();
|
|
}
|
|
|
|
void SetColumnFamily(uint32_t column_family_id) {
|
|
column_family_ = column_family_id;
|
|
}
|
|
uint32_t GetColumnFamily() const { return column_family_; }
|
|
|
|
// set column family ID by calling SetColumnFamily()
|
|
void AddColumnFamily(const std::string& name) {
|
|
assert(!is_column_family_drop_);
|
|
assert(!is_column_family_add_);
|
|
assert(NumEntries() == 0);
|
|
is_column_family_add_ = true;
|
|
column_family_name_ = name;
|
|
}
|
|
|
|
// set column family ID by calling SetColumnFamily()
|
|
void DropColumnFamily() {
|
|
assert(!is_column_family_drop_);
|
|
assert(!is_column_family_add_);
|
|
assert(NumEntries() == 0);
|
|
is_column_family_drop_ = true;
|
|
}
|
|
|
|
bool IsColumnFamilyManipulation() const {
|
|
return is_column_family_add_ || is_column_family_drop_;
|
|
}
|
|
|
|
bool IsColumnFamilyAdd() const { return is_column_family_add_; }
|
|
|
|
bool IsColumnFamilyDrop() const { return is_column_family_drop_; }
|
|
|
|
void MarkAtomicGroup(uint32_t remaining_entries) {
|
|
is_in_atomic_group_ = true;
|
|
remaining_entries_ = remaining_entries;
|
|
}
|
|
bool IsInAtomicGroup() const { return is_in_atomic_group_; }
|
|
uint32_t GetRemainingEntries() const { return remaining_entries_; }
|
|
|
|
bool HasFullHistoryTsLow() const { return !full_history_ts_low_.empty(); }
|
|
const std::string& GetFullHistoryTsLow() const {
|
|
assert(HasFullHistoryTsLow());
|
|
return full_history_ts_low_;
|
|
}
|
|
void SetFullHistoryTsLow(std::string full_history_ts_low) {
|
|
assert(!full_history_ts_low.empty());
|
|
full_history_ts_low_ = std::move(full_history_ts_low);
|
|
}
|
|
|
|
// return true on success.
|
|
// `ts_sz` is the size in bytes for the user-defined timestamp contained in
|
|
// a user key. This argument is optional because it's only required for
|
|
// encoding a `VersionEdit` with new SST files to add. It's used to handle the
|
|
// file boundaries: `smallest`, `largest` when
|
|
// `FileMetaData.user_defined_timestamps_persisted` is false. When reading
|
|
// the Manifest file, a mirroring change needed to handle
|
|
// file boundaries are not added to the `VersionEdit.DecodeFrom` function
|
|
// because timestamp size is not available at `VersionEdit` decoding time,
|
|
// it's instead added to `VersionEditHandler::OnNonCfOperation`.
|
|
bool EncodeTo(std::string* dst,
|
|
std::optional<size_t> ts_sz = std::nullopt) const;
|
|
Status DecodeFrom(const Slice& src);
|
|
|
|
std::string DebugString(bool hex_key = false) const;
|
|
std::string DebugJSON(int edit_num, bool hex_key = false) const;
|
|
|
|
private:
|
|
friend class ReactiveVersionSet;
|
|
friend class VersionEditHandlerBase;
|
|
friend class ListColumnFamiliesHandler;
|
|
friend class VersionEditHandler;
|
|
friend class VersionEditHandlerPointInTime;
|
|
friend class DumpManifestHandler;
|
|
friend class VersionSet;
|
|
friend class Version;
|
|
friend class AtomicGroupReadBuffer;
|
|
|
|
bool GetLevel(Slice* input, int* level, const char** msg);
|
|
|
|
const char* DecodeNewFile4From(Slice* input);
|
|
|
|
// Encode file boundaries `FileMetaData.smallest` and `FileMetaData.largest`.
|
|
// User-defined timestamps in the user key will be stripped if they shouldn't
|
|
// be persisted.
|
|
void EncodeFileBoundaries(std::string* dst, const FileMetaData& meta,
|
|
size_t ts_sz) const;
|
|
|
|
int max_level_ = 0;
|
|
std::string db_id_;
|
|
std::string comparator_;
|
|
uint64_t log_number_ = 0;
|
|
uint64_t prev_log_number_ = 0;
|
|
uint64_t next_file_number_ = 0;
|
|
uint32_t max_column_family_ = 0;
|
|
// The most recent WAL log number that is deleted
|
|
uint64_t min_log_number_to_keep_ = 0;
|
|
SequenceNumber last_sequence_ = 0;
|
|
bool has_db_id_ = false;
|
|
bool has_comparator_ = false;
|
|
bool has_log_number_ = false;
|
|
bool has_prev_log_number_ = false;
|
|
bool has_next_file_number_ = false;
|
|
bool has_max_column_family_ = false;
|
|
bool has_min_log_number_to_keep_ = false;
|
|
bool has_last_sequence_ = false;
|
|
bool has_persist_user_defined_timestamps_ = false;
|
|
|
|
// Compaction cursors for round-robin compaction policy
|
|
CompactCursors compact_cursors_;
|
|
|
|
DeletedFiles deleted_files_;
|
|
NewFiles new_files_;
|
|
|
|
BlobFileAdditions blob_file_additions_;
|
|
BlobFileGarbages blob_file_garbages_;
|
|
|
|
WalAdditions wal_additions_;
|
|
WalDeletion wal_deletion_;
|
|
|
|
// Each version edit record should have column_family_ set
|
|
// If it's not set, it is default (0)
|
|
uint32_t column_family_ = 0;
|
|
// a version edit can be either column_family add or
|
|
// column_family drop. If it's column family add,
|
|
// it also includes column family name.
|
|
bool is_column_family_drop_ = false;
|
|
bool is_column_family_add_ = false;
|
|
std::string column_family_name_;
|
|
|
|
bool is_in_atomic_group_ = false;
|
|
uint32_t remaining_entries_ = 0;
|
|
|
|
std::string full_history_ts_low_;
|
|
bool persist_user_defined_timestamps_ = true;
|
|
};
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|