2016-12-16 19:00:03 +00:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-15 23:03:42 +00:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2016-12-16 19:00:03 +00:00
|
|
|
#include <string>
|
|
|
|
#include <vector>
|
|
|
|
|
|
|
|
#include "db/db_test_util.h"
|
|
|
|
#include "db/forward_iterator.h"
|
|
|
|
#include "port/stack_trace.h"
|
2017-09-28 22:49:05 +00:00
|
|
|
#include "rocksdb/merge_operator.h"
|
2020-07-09 21:33:42 +00:00
|
|
|
#include "util/random.h"
|
2016-12-16 19:00:03 +00:00
|
|
|
#include "utilities/merge_operators.h"
|
2017-09-28 22:49:05 +00:00
|
|
|
#include "utilities/merge_operators/string_append/stringappend2.h"
|
2016-12-16 19:00:03 +00:00
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
2016-12-16 19:00:03 +00:00
|
|
|
|
2018-02-09 22:43:56 +00:00
|
|
|
class TestReadCallback : public ReadCallback {
|
|
|
|
public:
|
|
|
|
TestReadCallback(SnapshotChecker* snapshot_checker,
|
|
|
|
SequenceNumber snapshot_seq)
|
2019-04-02 21:43:03 +00:00
|
|
|
: ReadCallback(snapshot_seq),
|
|
|
|
snapshot_checker_(snapshot_checker),
|
|
|
|
snapshot_seq_(snapshot_seq) {}
|
2018-02-09 22:43:56 +00:00
|
|
|
|
2019-02-27 00:52:20 +00:00
|
|
|
bool IsVisibleFullCheck(SequenceNumber seq) override {
|
2019-01-16 17:48:10 +00:00
|
|
|
return snapshot_checker_->CheckInSnapshot(seq, snapshot_seq_) ==
|
|
|
|
SnapshotCheckerResult::kInSnapshot;
|
2018-02-09 22:43:56 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
SnapshotChecker* snapshot_checker_;
|
|
|
|
SequenceNumber snapshot_seq_;
|
|
|
|
};
|
|
|
|
|
2016-12-16 19:00:03 +00:00
|
|
|
// Test merge operator functionality.
|
|
|
|
class DBMergeOperatorTest : public DBTestBase {
|
|
|
|
public:
|
2020-08-18 01:41:20 +00:00
|
|
|
DBMergeOperatorTest()
|
2021-07-23 15:37:27 +00:00
|
|
|
: DBTestBase("db_merge_operator_test", /*env_do_fsync=*/false) {}
|
2018-02-09 22:43:56 +00:00
|
|
|
|
|
|
|
std::string GetWithReadCallback(SnapshotChecker* snapshot_checker,
|
|
|
|
const Slice& key,
|
|
|
|
const Snapshot* snapshot = nullptr) {
|
|
|
|
SequenceNumber seq = snapshot == nullptr ? db_->GetLatestSequenceNumber()
|
|
|
|
: snapshot->GetSequenceNumber();
|
|
|
|
TestReadCallback read_callback(snapshot_checker, seq);
|
|
|
|
ReadOptions read_opt;
|
|
|
|
read_opt.snapshot = snapshot;
|
|
|
|
PinnableSlice value;
|
New API to get all merge operands for a Key (#5604)
Summary:
This is a new API added to db.h to allow for fetching all merge operands associated with a Key. The main motivation for this API is to support use cases where doing a full online merge is not necessary as it is performance sensitive. Example use-cases:
1. Update subset of columns and read subset of columns -
Imagine a SQL Table, a row is encoded as a K/V pair (as it is done in MyRocks). If there are many columns and users only updated one of them, we can use merge operator to reduce write amplification. While users only read one or two columns in the read query, this feature can avoid a full merging of the whole row, and save some CPU.
2. Updating very few attributes in a value which is a JSON-like document -
Updating one attribute can be done efficiently using merge operator, while reading back one attribute can be done more efficiently if we don't need to do a full merge.
----------------------------------------------------------------------------------------------------
API :
Status GetMergeOperands(
const ReadOptions& options, ColumnFamilyHandle* column_family,
const Slice& key, PinnableSlice* merge_operands,
GetMergeOperandsOptions* get_merge_operands_options,
int* number_of_operands)
Example usage :
int size = 100;
int number_of_operands = 0;
std::vector<PinnableSlice> values(size);
GetMergeOperandsOptions merge_operands_info;
db_->GetMergeOperands(ReadOptions(), db_->DefaultColumnFamily(), "k1", values.data(), merge_operands_info, &number_of_operands);
Description :
Returns all the merge operands corresponding to the key. If the number of merge operands in DB is greater than merge_operands_options.expected_max_number_of_operands no merge operands are returned and status is Incomplete. Merge operands returned are in the order of insertion.
merge_operands-> Points to an array of at-least merge_operands_options.expected_max_number_of_operands and the caller is responsible for allocating it. If the status returned is Incomplete then number_of_operands will contain the total number of merge operands found in DB for key.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5604
Test Plan:
Added unit test and perf test in db_bench that can be run using the command:
./db_bench -benchmarks=getmergeoperands --merge_operator=sortlist
Differential Revision: D16657366
Pulled By: vjnadimpalli
fbshipit-source-id: 0faadd752351745224ee12d4ae9ef3cb529951bf
2019-08-06 21:22:34 +00:00
|
|
|
DBImpl::GetImplOptions get_impl_options;
|
|
|
|
get_impl_options.column_family = db_->DefaultColumnFamily();
|
|
|
|
get_impl_options.value = &value;
|
|
|
|
get_impl_options.callback = &read_callback;
|
|
|
|
Status s = dbfull()->GetImpl(read_opt, key, get_impl_options);
|
2018-02-09 22:43:56 +00:00
|
|
|
if (!s.ok()) {
|
|
|
|
return s.ToString();
|
|
|
|
}
|
|
|
|
return value.ToString();
|
|
|
|
}
|
2016-12-16 19:00:03 +00:00
|
|
|
};
|
|
|
|
|
2017-09-28 22:49:05 +00:00
|
|
|
TEST_F(DBMergeOperatorTest, LimitMergeOperands) {
|
|
|
|
class LimitedStringAppendMergeOp : public StringAppendTESTOperator {
|
|
|
|
public:
|
|
|
|
LimitedStringAppendMergeOp(int limit, char delim)
|
|
|
|
: StringAppendTESTOperator(delim), limit_(limit) {}
|
|
|
|
|
|
|
|
const char* Name() const override {
|
|
|
|
return "DBMergeOperatorTest::LimitedStringAppendMergeOp";
|
|
|
|
}
|
|
|
|
|
|
|
|
bool ShouldMerge(const std::vector<Slice>& operands) const override {
|
|
|
|
if (operands.size() > 0 && limit_ > 0 && operands.size() >= limit_) {
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
size_t limit_ = 0;
|
|
|
|
};
|
|
|
|
|
2023-06-16 04:12:39 +00:00
|
|
|
Options options = CurrentOptions();
|
2017-09-28 22:49:05 +00:00
|
|
|
options.create_if_missing = true;
|
|
|
|
// Use only the latest two merge operands.
|
2022-11-02 21:34:24 +00:00
|
|
|
options.merge_operator = std::make_shared<LimitedStringAppendMergeOp>(2, ',');
|
2017-09-28 22:49:05 +00:00
|
|
|
options.env = env_;
|
|
|
|
Reopen(options);
|
|
|
|
// All K1 values are in memtable.
|
|
|
|
ASSERT_OK(Merge("k1", "a"));
|
|
|
|
ASSERT_OK(Merge("k1", "b"));
|
|
|
|
ASSERT_OK(Merge("k1", "c"));
|
|
|
|
ASSERT_OK(Merge("k1", "d"));
|
|
|
|
std::string value;
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "k1", &value));
|
2017-09-28 22:49:05 +00:00
|
|
|
// Make sure that only the latest two merge operands are used. If this was
|
|
|
|
// not the case the value would be "a,b,c,d".
|
|
|
|
ASSERT_EQ(value, "c,d");
|
|
|
|
|
|
|
|
// All K2 values are flushed to L0 into a single file.
|
|
|
|
ASSERT_OK(Merge("k2", "a"));
|
|
|
|
ASSERT_OK(Merge("k2", "b"));
|
|
|
|
ASSERT_OK(Merge("k2", "c"));
|
|
|
|
ASSERT_OK(Merge("k2", "d"));
|
|
|
|
ASSERT_OK(Flush());
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "k2", &value));
|
2017-09-28 22:49:05 +00:00
|
|
|
ASSERT_EQ(value, "c,d");
|
|
|
|
|
|
|
|
// All K3 values are flushed and are in different files.
|
|
|
|
ASSERT_OK(Merge("k3", "ab"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(Merge("k3", "bc"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(Merge("k3", "cd"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_OK(Merge("k3", "de"));
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "k3", &value));
|
2017-09-28 22:49:05 +00:00
|
|
|
ASSERT_EQ(value, "cd,de");
|
|
|
|
|
|
|
|
// All K4 values are in different levels
|
|
|
|
ASSERT_OK(Merge("k4", "ab"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
MoveFilesToLevel(4);
|
|
|
|
ASSERT_OK(Merge("k4", "bc"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
MoveFilesToLevel(3);
|
|
|
|
ASSERT_OK(Merge("k4", "cd"));
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
MoveFilesToLevel(1);
|
|
|
|
ASSERT_OK(Merge("k4", "de"));
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "k4", &value));
|
2017-09-28 22:49:05 +00:00
|
|
|
ASSERT_EQ(value, "cd,de");
|
|
|
|
}
|
|
|
|
|
2016-12-16 19:00:03 +00:00
|
|
|
TEST_F(DBMergeOperatorTest, MergeErrorOnRead) {
|
2023-06-16 04:12:39 +00:00
|
|
|
Options options = CurrentOptions();
|
2016-12-16 19:00:03 +00:00
|
|
|
options.create_if_missing = true;
|
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
2017-03-13 16:41:30 +00:00
|
|
|
options.env = env_;
|
2016-12-16 19:00:03 +00:00
|
|
|
Reopen(options);
|
|
|
|
ASSERT_OK(Merge("k1", "v1"));
|
|
|
|
ASSERT_OK(Merge("k1", "corrupted"));
|
|
|
|
std::string value;
|
|
|
|
ASSERT_TRUE(db_->Get(ReadOptions(), "k1", &value).IsCorruption());
|
|
|
|
VerifyDBInternal({{"k1", "corrupted"}, {"k1", "v1"}});
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBMergeOperatorTest, MergeErrorOnWrite) {
|
2023-06-16 04:12:39 +00:00
|
|
|
Options options = CurrentOptions();
|
2016-12-16 19:00:03 +00:00
|
|
|
options.create_if_missing = true;
|
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
|
|
|
options.max_successive_merges = 3;
|
2017-03-13 16:41:30 +00:00
|
|
|
options.env = env_;
|
2016-12-16 19:00:03 +00:00
|
|
|
Reopen(options);
|
|
|
|
ASSERT_OK(Merge("k1", "v1"));
|
|
|
|
ASSERT_OK(Merge("k1", "v2"));
|
|
|
|
// Will trigger a merge when hitting max_successive_merges and the merge
|
|
|
|
// will fail. The delta will be inserted nevertheless.
|
|
|
|
ASSERT_OK(Merge("k1", "corrupted"));
|
|
|
|
// Data should stay unmerged after the error.
|
|
|
|
VerifyDBInternal({{"k1", "corrupted"}, {"k1", "v2"}, {"k1", "v1"}});
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_F(DBMergeOperatorTest, MergeErrorOnIteration) {
|
2023-06-16 04:12:39 +00:00
|
|
|
Options options = CurrentOptions();
|
2016-12-16 19:00:03 +00:00
|
|
|
options.create_if_missing = true;
|
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
2017-03-13 16:41:30 +00:00
|
|
|
options.env = env_;
|
2016-12-16 19:00:03 +00:00
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
ASSERT_OK(Merge("k1", "v1"));
|
|
|
|
ASSERT_OK(Merge("k1", "corrupted"));
|
|
|
|
ASSERT_OK(Put("k2", "v2"));
|
2017-12-06 19:54:45 +00:00
|
|
|
auto* iter = db_->NewIterator(ReadOptions());
|
|
|
|
iter->Seek("k1");
|
|
|
|
ASSERT_FALSE(iter->Valid());
|
|
|
|
ASSERT_TRUE(iter->status().IsCorruption());
|
|
|
|
delete iter;
|
|
|
|
iter = db_->NewIterator(ReadOptions());
|
|
|
|
iter->Seek("k2");
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_OK(iter->status());
|
|
|
|
iter->Prev();
|
|
|
|
ASSERT_FALSE(iter->Valid());
|
|
|
|
ASSERT_TRUE(iter->status().IsCorruption());
|
|
|
|
delete iter;
|
2016-12-16 19:00:03 +00:00
|
|
|
VerifyDBInternal({{"k1", "corrupted"}, {"k1", "v1"}, {"k2", "v2"}});
|
|
|
|
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
ASSERT_OK(Merge("k1", "v1"));
|
|
|
|
ASSERT_OK(Put("k2", "v2"));
|
|
|
|
ASSERT_OK(Merge("k2", "corrupted"));
|
2017-12-06 19:54:45 +00:00
|
|
|
iter = db_->NewIterator(ReadOptions());
|
|
|
|
iter->Seek("k1");
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_OK(iter->status());
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_FALSE(iter->Valid());
|
|
|
|
ASSERT_TRUE(iter->status().IsCorruption());
|
|
|
|
delete iter;
|
2016-12-16 19:00:03 +00:00
|
|
|
VerifyDBInternal({{"k1", "v1"}, {"k2", "corrupted"}, {"k2", "v2"}});
|
|
|
|
}
|
|
|
|
|
Add API to limit blast radius of merge operator failure (#11092)
Summary:
Prior to this PR, `FullMergeV2()` can only return `false` to indicate failure, which causes any operation invoking it to fail. During a compaction, such a failure causes the compaction to fail and causes the DB to irreversibly enter read-only mode. Some users asked for a way to allow the merge operator to fail without such widespread damage.
To limit the blast radius of merge operator failures, this PR introduces the `MergeOperationOutput::op_failure_scope` API. When unpopulated (`kDefault`) or set to `kTryMerge`, the merge operator failure handling is the same as before. When set to `kMustMerge`, merge operator failure still causes failure to operations that must merge (`Get()`, iterator, `MultiGet()`, etc.). However, under `kMustMerge`, flushes/compactions can survive merge operator failures by outputting the unmerged input operands.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/11092
Reviewed By: siying
Differential Revision: D42525673
Pulled By: ajkr
fbshipit-source-id: 951dc3bf190f86347dccf3381be967565cda52ee
2023-01-20 22:40:30 +00:00
|
|
|
|
|
|
|
TEST_F(DBMergeOperatorTest, MergeOperatorFailsWithMustMerge) {
|
|
|
|
// This is like a mini-stress test dedicated to `OpFailureScope::kMustMerge`.
|
|
|
|
// Some or most of it might be deleted upon adding that option to the actual
|
|
|
|
// stress test.
|
|
|
|
//
|
|
|
|
// "k0" and "k2" are stable (uncorrupted) keys before and after a corrupted
|
|
|
|
// key ("k1"). The outer loop (`i`) varies which write (`j`) to "k1" triggers
|
|
|
|
// the corruption. Inside that loop there are three cases:
|
|
|
|
//
|
|
|
|
// - Case 1: pure `Merge()`s
|
|
|
|
// - Case 2: `Merge()`s on top of a `Put()`
|
|
|
|
// - Case 3: `Merge()`s on top of a `Delete()`
|
|
|
|
//
|
|
|
|
// For each case we test query results before flush, after flush, and after
|
|
|
|
// compaction, as well as cleanup after deletion+compaction. The queries
|
|
|
|
// expect "k0" and "k2" to always be readable. "k1" is expected to be readable
|
|
|
|
// only by APIs that do not require merging, such as `GetMergeOperands()`.
|
|
|
|
const int kNumOperands = 3;
|
2023-06-16 04:12:39 +00:00
|
|
|
Options options = CurrentOptions();
|
Add API to limit blast radius of merge operator failure (#11092)
Summary:
Prior to this PR, `FullMergeV2()` can only return `false` to indicate failure, which causes any operation invoking it to fail. During a compaction, such a failure causes the compaction to fail and causes the DB to irreversibly enter read-only mode. Some users asked for a way to allow the merge operator to fail without such widespread damage.
To limit the blast radius of merge operator failures, this PR introduces the `MergeOperationOutput::op_failure_scope` API. When unpopulated (`kDefault`) or set to `kTryMerge`, the merge operator failure handling is the same as before. When set to `kMustMerge`, merge operator failure still causes failure to operations that must merge (`Get()`, iterator, `MultiGet()`, etc.). However, under `kMustMerge`, flushes/compactions can survive merge operator failures by outputting the unmerged input operands.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/11092
Reviewed By: siying
Differential Revision: D42525673
Pulled By: ajkr
fbshipit-source-id: 951dc3bf190f86347dccf3381be967565cda52ee
2023-01-20 22:40:30 +00:00
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
|
|
|
options.env = env_;
|
|
|
|
Reopen(options);
|
|
|
|
|
|
|
|
for (int i = 0; i < kNumOperands; ++i) {
|
|
|
|
auto check_query = [&]() {
|
|
|
|
{
|
|
|
|
std::string value;
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "k0", &value));
|
2023-02-17 18:58:46 +00:00
|
|
|
Status s = db_->Get(ReadOptions(), "k1", &value);
|
|
|
|
ASSERT_TRUE(s.IsCorruption());
|
|
|
|
ASSERT_EQ(Status::SubCode::kMergeOperatorFailed, s.subcode());
|
Add API to limit blast radius of merge operator failure (#11092)
Summary:
Prior to this PR, `FullMergeV2()` can only return `false` to indicate failure, which causes any operation invoking it to fail. During a compaction, such a failure causes the compaction to fail and causes the DB to irreversibly enter read-only mode. Some users asked for a way to allow the merge operator to fail without such widespread damage.
To limit the blast radius of merge operator failures, this PR introduces the `MergeOperationOutput::op_failure_scope` API. When unpopulated (`kDefault`) or set to `kTryMerge`, the merge operator failure handling is the same as before. When set to `kMustMerge`, merge operator failure still causes failure to operations that must merge (`Get()`, iterator, `MultiGet()`, etc.). However, under `kMustMerge`, flushes/compactions can survive merge operator failures by outputting the unmerged input operands.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/11092
Reviewed By: siying
Differential Revision: D42525673
Pulled By: ajkr
fbshipit-source-id: 951dc3bf190f86347dccf3381be967565cda52ee
2023-01-20 22:40:30 +00:00
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "k2", &value));
|
|
|
|
}
|
|
|
|
|
|
|
|
{
|
|
|
|
std::unique_ptr<Iterator> iter;
|
|
|
|
iter.reset(db_->NewIterator(ReadOptions()));
|
|
|
|
iter->SeekToFirst();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ("k0", iter->key());
|
|
|
|
iter->Next();
|
|
|
|
ASSERT_TRUE(iter->status().IsCorruption());
|
2023-02-17 18:58:46 +00:00
|
|
|
ASSERT_EQ(Status::SubCode::kMergeOperatorFailed,
|
|
|
|
iter->status().subcode());
|
Add API to limit blast radius of merge operator failure (#11092)
Summary:
Prior to this PR, `FullMergeV2()` can only return `false` to indicate failure, which causes any operation invoking it to fail. During a compaction, such a failure causes the compaction to fail and causes the DB to irreversibly enter read-only mode. Some users asked for a way to allow the merge operator to fail without such widespread damage.
To limit the blast radius of merge operator failures, this PR introduces the `MergeOperationOutput::op_failure_scope` API. When unpopulated (`kDefault`) or set to `kTryMerge`, the merge operator failure handling is the same as before. When set to `kMustMerge`, merge operator failure still causes failure to operations that must merge (`Get()`, iterator, `MultiGet()`, etc.). However, under `kMustMerge`, flushes/compactions can survive merge operator failures by outputting the unmerged input operands.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/11092
Reviewed By: siying
Differential Revision: D42525673
Pulled By: ajkr
fbshipit-source-id: 951dc3bf190f86347dccf3381be967565cda52ee
2023-01-20 22:40:30 +00:00
|
|
|
|
|
|
|
iter->SeekToLast();
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ("k2", iter->key());
|
|
|
|
iter->Prev();
|
|
|
|
ASSERT_TRUE(iter->status().IsCorruption());
|
|
|
|
|
|
|
|
iter->Seek("k2");
|
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
ASSERT_EQ("k2", iter->key());
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<PinnableSlice> values(kNumOperands);
|
|
|
|
GetMergeOperandsOptions merge_operands_info;
|
|
|
|
merge_operands_info.expected_max_number_of_operands = kNumOperands;
|
|
|
|
int num_operands_found = 0;
|
|
|
|
ASSERT_OK(db_->GetMergeOperands(ReadOptions(), db_->DefaultColumnFamily(),
|
|
|
|
"k1", values.data(), &merge_operands_info,
|
|
|
|
&num_operands_found));
|
|
|
|
ASSERT_EQ(kNumOperands, num_operands_found);
|
|
|
|
for (int j = 0; j < num_operands_found; ++j) {
|
|
|
|
if (i == j) {
|
|
|
|
ASSERT_EQ(values[j], "corrupted_must_merge");
|
|
|
|
} else {
|
|
|
|
ASSERT_EQ(values[j], "ok");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
ASSERT_OK(Put("k0", "val"));
|
|
|
|
ASSERT_OK(Put("k2", "val"));
|
|
|
|
|
|
|
|
// Case 1
|
|
|
|
for (int j = 0; j < kNumOperands; ++j) {
|
|
|
|
if (j == i) {
|
|
|
|
ASSERT_OK(Merge("k1", "corrupted_must_merge"));
|
|
|
|
} else {
|
|
|
|
ASSERT_OK(Merge("k1", "ok"));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
check_query();
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
check_query();
|
|
|
|
{
|
|
|
|
CompactRangeOptions cro;
|
|
|
|
cro.bottommost_level_compaction =
|
|
|
|
BottommostLevelCompaction::kForceOptimized;
|
|
|
|
ASSERT_OK(db_->CompactRange(cro, nullptr, nullptr));
|
|
|
|
}
|
|
|
|
check_query();
|
|
|
|
|
|
|
|
// Case 2
|
|
|
|
for (int j = 0; j < kNumOperands; ++j) {
|
|
|
|
Slice val;
|
|
|
|
if (j == i) {
|
|
|
|
val = "corrupted_must_merge";
|
|
|
|
} else {
|
|
|
|
val = "ok";
|
|
|
|
}
|
|
|
|
if (j == 0) {
|
|
|
|
ASSERT_OK(Put("k1", val));
|
|
|
|
} else {
|
|
|
|
ASSERT_OK(Merge("k1", val));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
check_query();
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
check_query();
|
|
|
|
{
|
|
|
|
CompactRangeOptions cro;
|
|
|
|
cro.bottommost_level_compaction =
|
|
|
|
BottommostLevelCompaction::kForceOptimized;
|
|
|
|
ASSERT_OK(db_->CompactRange(cro, nullptr, nullptr));
|
|
|
|
}
|
|
|
|
check_query();
|
|
|
|
|
|
|
|
// Case 3
|
|
|
|
ASSERT_OK(Delete("k1"));
|
|
|
|
for (int j = 0; j < kNumOperands; ++j) {
|
|
|
|
if (i == j) {
|
|
|
|
ASSERT_OK(Merge("k1", "corrupted_must_merge"));
|
|
|
|
} else {
|
|
|
|
ASSERT_OK(Merge("k1", "ok"));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
check_query();
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
check_query();
|
|
|
|
{
|
|
|
|
CompactRangeOptions cro;
|
|
|
|
cro.bottommost_level_compaction =
|
|
|
|
BottommostLevelCompaction::kForceOptimized;
|
|
|
|
ASSERT_OK(db_->CompactRange(cro, nullptr, nullptr));
|
|
|
|
}
|
|
|
|
check_query();
|
|
|
|
|
|
|
|
// Verify obsolete data removal still happens
|
|
|
|
ASSERT_OK(Delete("k0"));
|
|
|
|
ASSERT_OK(Delete("k1"));
|
|
|
|
ASSERT_OK(Delete("k2"));
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get("k0"));
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get("k1"));
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get("k2"));
|
|
|
|
CompactRangeOptions cro;
|
|
|
|
cro.bottommost_level_compaction =
|
|
|
|
BottommostLevelCompaction::kForceOptimized;
|
|
|
|
ASSERT_OK(db_->CompactRange(cro, nullptr, nullptr));
|
|
|
|
ASSERT_EQ("", FilesPerLevel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-04-07 17:06:03 +00:00
|
|
|
TEST_F(DBMergeOperatorTest, DataBlockBinaryAndHash) {
|
|
|
|
// Basic test to check that merge operator works with data block index type
|
|
|
|
// DataBlockBinaryAndHash.
|
2023-06-16 04:12:39 +00:00
|
|
|
Options options = CurrentOptions();
|
2023-04-07 17:06:03 +00:00
|
|
|
options.create_if_missing = true;
|
|
|
|
options.merge_operator.reset(new TestPutOperator());
|
|
|
|
options.env = env_;
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.block_restart_interval = 16;
|
|
|
|
table_options.data_block_index_type =
|
|
|
|
BlockBasedTableOptions::DataBlockIndexType::kDataBlockBinaryAndHash;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
const int kNumKeys = 100;
|
|
|
|
for (int i = 0; i < kNumKeys; ++i) {
|
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), Key(i), std::to_string(i)));
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
std::string value;
|
|
|
|
for (int i = 0; i < kNumKeys; ++i) {
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), Key(i), &value));
|
|
|
|
ASSERT_EQ(std::to_string(i), value);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<const Snapshot*> snapshots;
|
|
|
|
for (int i = 0; i < kNumKeys; ++i) {
|
|
|
|
ASSERT_OK(db_->Delete(WriteOptions(), Key(i)));
|
|
|
|
for (int j = 0; j < 3; ++j) {
|
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), Key(i), std::to_string(i * 3 + j)));
|
|
|
|
snapshots.push_back(db_->GetSnapshot());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
for (int i = 0; i < kNumKeys; ++i) {
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), Key(i), &value));
|
|
|
|
ASSERT_EQ(std::to_string(i * 3 + 2), value);
|
|
|
|
}
|
|
|
|
for (auto snapshot : snapshots) {
|
|
|
|
db_->ReleaseSnapshot(snapshot);
|
|
|
|
}
|
|
|
|
}
|
Add API to limit blast radius of merge operator failure (#11092)
Summary:
Prior to this PR, `FullMergeV2()` can only return `false` to indicate failure, which causes any operation invoking it to fail. During a compaction, such a failure causes the compaction to fail and causes the DB to irreversibly enter read-only mode. Some users asked for a way to allow the merge operator to fail without such widespread damage.
To limit the blast radius of merge operator failures, this PR introduces the `MergeOperationOutput::op_failure_scope` API. When unpopulated (`kDefault`) or set to `kTryMerge`, the merge operator failure handling is the same as before. When set to `kMustMerge`, merge operator failure still causes failure to operations that must merge (`Get()`, iterator, `MultiGet()`, etc.). However, under `kMustMerge`, flushes/compactions can survive merge operator failures by outputting the unmerged input operands.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/11092
Reviewed By: siying
Differential Revision: D42525673
Pulled By: ajkr
fbshipit-source-id: 951dc3bf190f86347dccf3381be967565cda52ee
2023-01-20 22:40:30 +00:00
|
|
|
|
2017-04-11 23:11:27 +00:00
|
|
|
class MergeOperatorPinningTest : public DBMergeOperatorTest,
|
|
|
|
public testing::WithParamInterface<bool> {
|
|
|
|
public:
|
|
|
|
MergeOperatorPinningTest() { disable_block_cache_ = GetParam(); }
|
|
|
|
|
|
|
|
bool disable_block_cache_;
|
|
|
|
};
|
|
|
|
|
2020-06-03 22:53:09 +00:00
|
|
|
INSTANTIATE_TEST_CASE_P(MergeOperatorPinningTest, MergeOperatorPinningTest,
|
|
|
|
::testing::Bool());
|
2017-04-11 23:11:27 +00:00
|
|
|
|
|
|
|
TEST_P(MergeOperatorPinningTest, OperandsMultiBlocks) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.block_size = 1; // every block will contain one entry
|
|
|
|
table_options.no_block_cache = disable_block_cache_;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
options.merge_operator = MergeOperators::CreateStringAppendTESTOperator();
|
|
|
|
options.level0_slowdown_writes_trigger = (1 << 30);
|
|
|
|
options.level0_stop_writes_trigger = (1 << 30);
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
const int kKeysPerFile = 10;
|
|
|
|
const int kOperandsPerKeyPerFile = 7;
|
|
|
|
const int kOperandSize = 100;
|
|
|
|
// Filse to write in L0 before compacting to lower level
|
|
|
|
const int kFilesPerLevel = 3;
|
|
|
|
|
|
|
|
Random rnd(301);
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
int batch_num = 1;
|
|
|
|
int lvl_to_fill = 4;
|
|
|
|
int key_id = 0;
|
|
|
|
while (true) {
|
|
|
|
for (int j = 0; j < kKeysPerFile; j++) {
|
|
|
|
std::string key = Key(key_id % 35);
|
|
|
|
key_id++;
|
|
|
|
for (int k = 0; k < kOperandsPerKeyPerFile; k++) {
|
2020-07-09 21:33:42 +00:00
|
|
|
std::string val = rnd.RandomString(kOperandSize);
|
2017-04-11 23:11:27 +00:00
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), key, val));
|
|
|
|
if (true_data[key].size() == 0) {
|
|
|
|
true_data[key] = val;
|
|
|
|
} else {
|
|
|
|
true_data[key] += "," + val;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (lvl_to_fill == -1) {
|
|
|
|
// Keep last batch in memtable and stop
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
if (batch_num % kFilesPerLevel == 0) {
|
|
|
|
if (lvl_to_fill != 0) {
|
|
|
|
MoveFilesToLevel(lvl_to_fill);
|
|
|
|
}
|
|
|
|
lvl_to_fill--;
|
|
|
|
}
|
|
|
|
batch_num++;
|
|
|
|
}
|
|
|
|
|
|
|
|
// 3 L0 files
|
|
|
|
// 1 L1 file
|
|
|
|
// 3 L2 files
|
|
|
|
// 1 L3 file
|
|
|
|
// 3 L4 Files
|
|
|
|
ASSERT_EQ(FilesPerLevel(), "3,1,3,1,3");
|
|
|
|
|
|
|
|
VerifyDBFromMap(true_data);
|
|
|
|
}
|
|
|
|
|
|
|
|
class MergeOperatorHook : public MergeOperator {
|
|
|
|
public:
|
|
|
|
explicit MergeOperatorHook(std::shared_ptr<MergeOperator> _merge_op)
|
|
|
|
: merge_op_(_merge_op) {}
|
|
|
|
|
2019-02-14 21:52:47 +00:00
|
|
|
bool FullMergeV2(const MergeOperationInput& merge_in,
|
|
|
|
MergeOperationOutput* merge_out) const override {
|
2017-04-11 23:11:27 +00:00
|
|
|
before_merge_();
|
|
|
|
bool res = merge_op_->FullMergeV2(merge_in, merge_out);
|
|
|
|
after_merge_();
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2019-02-14 21:52:47 +00:00
|
|
|
const char* Name() const override { return merge_op_->Name(); }
|
2017-04-11 23:11:27 +00:00
|
|
|
|
|
|
|
std::shared_ptr<MergeOperator> merge_op_;
|
|
|
|
std::function<void()> before_merge_ = []() {};
|
|
|
|
std::function<void()> after_merge_ = []() {};
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_P(MergeOperatorPinningTest, EvictCacheBeforeMerge) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
|
|
|
|
auto merge_hook =
|
|
|
|
std::make_shared<MergeOperatorHook>(MergeOperators::CreateMaxOperator());
|
|
|
|
options.merge_operator = merge_hook;
|
|
|
|
options.disable_auto_compactions = true;
|
|
|
|
options.level0_slowdown_writes_trigger = (1 << 30);
|
|
|
|
options.level0_stop_writes_trigger = (1 << 30);
|
|
|
|
options.max_open_files = 20;
|
|
|
|
BlockBasedTableOptions bbto;
|
|
|
|
bbto.no_block_cache = disable_block_cache_;
|
|
|
|
if (bbto.no_block_cache == false) {
|
|
|
|
bbto.block_cache = NewLRUCache(64 * 1024 * 1024);
|
|
|
|
} else {
|
|
|
|
bbto.block_cache = nullptr;
|
|
|
|
}
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(bbto));
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
const int kNumOperands = 30;
|
|
|
|
const int kNumKeys = 1000;
|
|
|
|
const int kOperandSize = 100;
|
|
|
|
Random rnd(301);
|
|
|
|
|
|
|
|
// 1000 keys every key have 30 operands, every operand is in a different file
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
for (int i = 0; i < kNumOperands; i++) {
|
|
|
|
for (int j = 0; j < kNumKeys; j++) {
|
|
|
|
std::string k = Key(j);
|
2020-07-09 21:33:42 +00:00
|
|
|
std::string v = rnd.RandomString(kOperandSize);
|
2017-04-11 23:11:27 +00:00
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), k, v));
|
|
|
|
|
|
|
|
true_data[k] = std::max(true_data[k], v);
|
|
|
|
}
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<uint64_t> file_numbers = ListTableFiles(env_, dbname_);
|
|
|
|
ASSERT_EQ(file_numbers.size(), kNumOperands);
|
|
|
|
int merge_cnt = 0;
|
|
|
|
|
|
|
|
// Code executed before merge operation
|
|
|
|
merge_hook->before_merge_ = [&]() {
|
|
|
|
// Evict all tables from cache before every merge operation
|
2020-12-08 23:53:59 +00:00
|
|
|
auto* table_cache = dbfull()->TEST_table_cache();
|
2017-04-11 23:11:27 +00:00
|
|
|
for (uint64_t num : file_numbers) {
|
2020-12-08 23:53:59 +00:00
|
|
|
TableCache::Evict(table_cache, num);
|
2017-04-11 23:11:27 +00:00
|
|
|
}
|
|
|
|
// Decrease cache capacity to force all unrefed blocks to be evicted
|
|
|
|
if (bbto.block_cache) {
|
|
|
|
bbto.block_cache->SetCapacity(1);
|
|
|
|
}
|
|
|
|
merge_cnt++;
|
|
|
|
};
|
|
|
|
|
|
|
|
// Code executed after merge operation
|
|
|
|
merge_hook->after_merge_ = [&]() {
|
|
|
|
// Increase capacity again after doing the merge
|
|
|
|
if (bbto.block_cache) {
|
|
|
|
bbto.block_cache->SetCapacity(64 * 1024 * 1024);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
size_t total_reads;
|
|
|
|
VerifyDBFromMap(true_data, &total_reads);
|
|
|
|
ASSERT_EQ(merge_cnt, total_reads);
|
|
|
|
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
2017-04-11 23:11:27 +00:00
|
|
|
|
|
|
|
VerifyDBFromMap(true_data, &total_reads);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(MergeOperatorPinningTest, TailingIterator) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.merge_operator = MergeOperators::CreateMaxOperator();
|
|
|
|
BlockBasedTableOptions bbto;
|
|
|
|
bbto.no_block_cache = disable_block_cache_;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(bbto));
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
const int kNumOperands = 100;
|
|
|
|
const int kNumWrites = 100000;
|
|
|
|
|
|
|
|
std::function<void()> writer_func = [&]() {
|
|
|
|
int k = 0;
|
|
|
|
for (int i = 0; i < kNumWrites; i++) {
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), Key(k), Key(k)));
|
2017-04-11 23:11:27 +00:00
|
|
|
|
|
|
|
if (i && i % kNumOperands == 0) {
|
|
|
|
k++;
|
|
|
|
}
|
|
|
|
if (i && i % 127 == 0) {
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
}
|
|
|
|
if (i && i % 317 == 0) {
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
std::function<void()> reader_func = [&]() {
|
|
|
|
ReadOptions ro;
|
|
|
|
ro.tailing = true;
|
|
|
|
Iterator* iter = db_->NewIterator(ro);
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(iter->status());
|
2017-04-11 23:11:27 +00:00
|
|
|
iter->SeekToFirst();
|
|
|
|
for (int i = 0; i < (kNumWrites / kNumOperands); i++) {
|
|
|
|
while (!iter->Valid()) {
|
|
|
|
// wait for the key to be written
|
|
|
|
env_->SleepForMicroseconds(100);
|
|
|
|
iter->Seek(Key(i));
|
|
|
|
}
|
|
|
|
ASSERT_EQ(iter->key(), Key(i));
|
|
|
|
ASSERT_EQ(iter->value(), Key(i));
|
|
|
|
|
|
|
|
iter->Next();
|
|
|
|
}
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(iter->status());
|
2017-04-11 23:11:27 +00:00
|
|
|
|
|
|
|
delete iter;
|
|
|
|
};
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::port::Thread writer_thread(writer_func);
|
|
|
|
ROCKSDB_NAMESPACE::port::Thread reader_thread(reader_func);
|
2017-04-11 23:11:27 +00:00
|
|
|
|
|
|
|
writer_thread.join();
|
|
|
|
reader_thread.join();
|
|
|
|
}
|
2018-02-03 05:16:26 +00:00
|
|
|
|
|
|
|
TEST_F(DBMergeOperatorTest, TailingIteratorMemtableUnrefedBySomeoneElse) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.merge_operator = MergeOperators::CreateStringAppendOperator();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
// Overview of the test:
|
|
|
|
// * There are two merge operands for the same key: one in an sst file,
|
|
|
|
// another in a memtable.
|
|
|
|
// * Seek a tailing iterator to this key.
|
|
|
|
// * As part of the seek, the iterator will:
|
|
|
|
// (a) first visit the operand in the memtable and tell ForwardIterator
|
|
|
|
// to pin this operand, then
|
|
|
|
// (b) move on to the operand in the sst file, then pass both operands
|
|
|
|
// to merge operator.
|
|
|
|
// * The memtable may get flushed and unreferenced by another thread between
|
|
|
|
// (a) and (b). The test simulates it by flushing the memtable inside a
|
|
|
|
// SyncPoint callback located between (a) and (b).
|
|
|
|
// * In this case it's ForwardIterator's responsibility to keep the memtable
|
|
|
|
// pinned until (b) is complete. There used to be a bug causing
|
|
|
|
// ForwardIterator to not pin it in some circumstances. This test
|
|
|
|
// reproduces it.
|
|
|
|
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), "key", "sst"));
|
|
|
|
ASSERT_OK(db_->Flush(FlushOptions())); // Switch to SuperVersion A
|
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), "key", "memtable"));
|
2018-02-03 05:16:26 +00:00
|
|
|
|
|
|
|
// Pin SuperVersion A
|
|
|
|
std::unique_ptr<Iterator> someone_else(db_->NewIterator(ReadOptions()));
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(someone_else->status());
|
2018-02-03 05:16:26 +00:00
|
|
|
|
|
|
|
bool pushed_first_operand = false;
|
|
|
|
bool stepped_to_next_operand = false;
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
|
2018-02-03 05:16:26 +00:00
|
|
|
"DBIter::MergeValuesNewToOld:PushedFirstOperand", [&](void*) {
|
|
|
|
EXPECT_FALSE(pushed_first_operand);
|
|
|
|
pushed_first_operand = true;
|
2020-12-08 23:53:59 +00:00
|
|
|
EXPECT_OK(db_->Flush(FlushOptions())); // Switch to SuperVersion B
|
2018-02-03 05:16:26 +00:00
|
|
|
});
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
|
2018-02-03 05:16:26 +00:00
|
|
|
"DBIter::MergeValuesNewToOld:SteppedToNextOperand", [&](void*) {
|
|
|
|
EXPECT_FALSE(stepped_to_next_operand);
|
|
|
|
stepped_to_next_operand = true;
|
2022-11-02 21:34:24 +00:00
|
|
|
someone_else.reset(); // Unpin SuperVersion A
|
2018-02-03 05:16:26 +00:00
|
|
|
});
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
2018-02-03 05:16:26 +00:00
|
|
|
|
|
|
|
ReadOptions ro;
|
|
|
|
ro.tailing = true;
|
|
|
|
std::unique_ptr<Iterator> iter(db_->NewIterator(ro));
|
|
|
|
iter->Seek("key");
|
|
|
|
|
2020-12-08 23:53:59 +00:00
|
|
|
ASSERT_OK(iter->status());
|
2018-02-03 05:16:26 +00:00
|
|
|
ASSERT_TRUE(iter->Valid());
|
|
|
|
EXPECT_EQ(std::string("sst,memtable"), iter->value().ToString());
|
|
|
|
EXPECT_TRUE(pushed_first_operand);
|
|
|
|
EXPECT_TRUE(stepped_to_next_operand);
|
|
|
|
}
|
2017-04-11 23:11:27 +00:00
|
|
|
|
2018-02-09 22:43:56 +00:00
|
|
|
TEST_F(DBMergeOperatorTest, SnapshotCheckerAndReadCallback) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.merge_operator = MergeOperators::CreateStringAppendOperator();
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
class TestSnapshotChecker : public SnapshotChecker {
|
2019-01-16 17:48:10 +00:00
|
|
|
public:
|
|
|
|
SnapshotCheckerResult CheckInSnapshot(
|
|
|
|
SequenceNumber seq, SequenceNumber snapshot_seq) const override {
|
|
|
|
return IsInSnapshot(seq, snapshot_seq)
|
|
|
|
? SnapshotCheckerResult::kInSnapshot
|
|
|
|
: SnapshotCheckerResult::kNotInSnapshot;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool IsInSnapshot(SequenceNumber seq, SequenceNumber snapshot_seq) const {
|
2018-02-09 22:43:56 +00:00
|
|
|
switch (snapshot_seq) {
|
|
|
|
case 0:
|
|
|
|
return seq == 0;
|
|
|
|
case 1:
|
|
|
|
return seq <= 1;
|
|
|
|
case 2:
|
|
|
|
// seq = 2 not visible to snapshot with seq = 2
|
|
|
|
return seq <= 1;
|
|
|
|
case 3:
|
|
|
|
return seq <= 3;
|
|
|
|
case 4:
|
|
|
|
// seq = 4 not visible to snpahost with seq = 4
|
|
|
|
return seq <= 3;
|
|
|
|
default:
|
|
|
|
// seq >=4 is uncommitted
|
|
|
|
return seq <= 4;
|
|
|
|
};
|
|
|
|
}
|
|
|
|
};
|
|
|
|
TestSnapshotChecker* snapshot_checker = new TestSnapshotChecker();
|
|
|
|
dbfull()->SetSnapshotChecker(snapshot_checker);
|
|
|
|
|
|
|
|
std::string value;
|
|
|
|
ASSERT_OK(Merge("foo", "v1"));
|
|
|
|
ASSERT_EQ(1, db_->GetLatestSequenceNumber());
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
ASSERT_OK(Merge("foo", "v2"));
|
|
|
|
ASSERT_EQ(2, db_->GetLatestSequenceNumber());
|
|
|
|
// v2 is not visible to latest snapshot, which has seq = 2.
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
// Take a snapshot with seq = 2.
|
|
|
|
const Snapshot* snapshot1 = db_->GetSnapshot();
|
|
|
|
ASSERT_EQ(2, snapshot1->GetSequenceNumber());
|
|
|
|
// v2 is not visible to snapshot1, which has seq = 2
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo", snapshot1));
|
|
|
|
|
|
|
|
// Verify flush doesn't alter the result.
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo", snapshot1));
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
|
|
|
|
ASSERT_OK(Merge("foo", "v3"));
|
|
|
|
ASSERT_EQ(3, db_->GetLatestSequenceNumber());
|
|
|
|
ASSERT_EQ("v1,v2,v3", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
ASSERT_OK(Merge("foo", "v4"));
|
|
|
|
ASSERT_EQ(4, db_->GetLatestSequenceNumber());
|
|
|
|
// v4 is not visible to latest snapshot, which has seq = 4.
|
|
|
|
ASSERT_EQ("v1,v2,v3", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
const Snapshot* snapshot2 = db_->GetSnapshot();
|
|
|
|
ASSERT_EQ(4, snapshot2->GetSequenceNumber());
|
|
|
|
// v4 is not visible to snapshot2, which has seq = 4.
|
|
|
|
ASSERT_EQ("v1,v2,v3",
|
|
|
|
GetWithReadCallback(snapshot_checker, "foo", snapshot2));
|
|
|
|
|
|
|
|
// Verify flush doesn't alter the result.
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo", snapshot1));
|
|
|
|
ASSERT_EQ("v1,v2,v3",
|
|
|
|
GetWithReadCallback(snapshot_checker, "foo", snapshot2));
|
|
|
|
ASSERT_EQ("v1,v2,v3", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
|
|
|
|
ASSERT_OK(Merge("foo", "v5"));
|
|
|
|
ASSERT_EQ(5, db_->GetLatestSequenceNumber());
|
|
|
|
// v5 is uncommitted
|
|
|
|
ASSERT_EQ("v1,v2,v3,v4", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
|
|
|
|
// full manual compaction.
|
|
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
|
|
|
|
// Verify compaction doesn't alter the result.
|
|
|
|
ASSERT_EQ("v1", GetWithReadCallback(snapshot_checker, "foo", snapshot1));
|
|
|
|
ASSERT_EQ("v1,v2,v3",
|
|
|
|
GetWithReadCallback(snapshot_checker, "foo", snapshot2));
|
|
|
|
ASSERT_EQ("v1,v2,v3,v4", GetWithReadCallback(snapshot_checker, "foo"));
|
|
|
|
|
|
|
|
db_->ReleaseSnapshot(snapshot1);
|
|
|
|
db_->ReleaseSnapshot(snapshot2);
|
|
|
|
}
|
|
|
|
|
2019-07-31 22:16:01 +00:00
|
|
|
class PerConfigMergeOperatorPinningTest
|
|
|
|
: public DBMergeOperatorTest,
|
|
|
|
public testing::WithParamInterface<std::tuple<bool, int>> {
|
|
|
|
public:
|
|
|
|
PerConfigMergeOperatorPinningTest() {
|
|
|
|
std::tie(disable_block_cache_, option_config_) = GetParam();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool disable_block_cache_;
|
|
|
|
};
|
|
|
|
|
2020-06-03 22:53:09 +00:00
|
|
|
INSTANTIATE_TEST_CASE_P(
|
2019-07-31 22:16:01 +00:00
|
|
|
MergeOperatorPinningTest, PerConfigMergeOperatorPinningTest,
|
|
|
|
::testing::Combine(::testing::Bool(),
|
|
|
|
::testing::Range(static_cast<int>(DBTestBase::kDefault),
|
|
|
|
static_cast<int>(DBTestBase::kEnd))));
|
|
|
|
|
|
|
|
TEST_P(PerConfigMergeOperatorPinningTest, Randomized) {
|
|
|
|
if (ShouldSkipOptions(option_config_, kSkipMergePut)) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
options.merge_operator = MergeOperators::CreateMaxOperator();
|
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.no_block_cache = disable_block_cache_;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
|
|
|
|
Random rnd(301);
|
|
|
|
std::map<std::string, std::string> true_data;
|
|
|
|
|
|
|
|
const int kTotalMerges = 5000;
|
|
|
|
// Every key gets ~10 operands
|
|
|
|
const int kKeyRange = kTotalMerges / 10;
|
|
|
|
const int kOperandSize = 20;
|
|
|
|
const int kNumPutBefore = kKeyRange / 10; // 10% value
|
|
|
|
const int kNumPutAfter = kKeyRange / 10; // 10% overwrite
|
|
|
|
const int kNumDelete = kKeyRange / 10; // 10% delete
|
|
|
|
|
|
|
|
// kNumPutBefore keys will have base values
|
|
|
|
for (int i = 0; i < kNumPutBefore; i++) {
|
|
|
|
std::string key = Key(rnd.Next() % kKeyRange);
|
2020-07-09 21:33:42 +00:00
|
|
|
std::string value = rnd.RandomString(kOperandSize);
|
2019-07-31 22:16:01 +00:00
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), key, value));
|
|
|
|
|
|
|
|
true_data[key] = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Do kTotalMerges merges
|
|
|
|
for (int i = 0; i < kTotalMerges; i++) {
|
|
|
|
std::string key = Key(rnd.Next() % kKeyRange);
|
2020-07-09 21:33:42 +00:00
|
|
|
std::string value = rnd.RandomString(kOperandSize);
|
2019-07-31 22:16:01 +00:00
|
|
|
ASSERT_OK(db_->Merge(WriteOptions(), key, value));
|
|
|
|
|
|
|
|
if (true_data[key] < value) {
|
|
|
|
true_data[key] = value;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Overwrite random kNumPutAfter keys
|
|
|
|
for (int i = 0; i < kNumPutAfter; i++) {
|
|
|
|
std::string key = Key(rnd.Next() % kKeyRange);
|
2020-07-09 21:33:42 +00:00
|
|
|
std::string value = rnd.RandomString(kOperandSize);
|
2019-07-31 22:16:01 +00:00
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), key, value));
|
|
|
|
|
|
|
|
true_data[key] = value;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete random kNumDelete keys
|
|
|
|
for (int i = 0; i < kNumDelete; i++) {
|
|
|
|
std::string key = Key(rnd.Next() % kKeyRange);
|
|
|
|
ASSERT_OK(db_->Delete(WriteOptions(), key));
|
|
|
|
|
|
|
|
true_data.erase(key);
|
|
|
|
}
|
|
|
|
|
|
|
|
VerifyDBFromMap(true_data);
|
|
|
|
}
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
2016-12-16 19:00:03 +00:00
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
|
2016-12-16 19:00:03 +00:00
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
}
|