mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-30 13:41:46 +00:00
12f1137355
Summary: Introduces and uses a SystemClock class to RocksDB. This class contains the time-related functions of an Env and these functions can be redirected from the Env to the SystemClock. Many of the places that used an Env (Timer, PerfStepTimer, RepeatableThread, RateLimiter, WriteController) for time-related functions have been changed to use SystemClock instead. There are likely more places that can be changed, but this is a start to show what can/should be done. Over time it would be nice to migrate most (if not all) of the uses of the time functions from the Env to the SystemClock. There are several Env classes that implement these functions. Most of these have not been converted yet to SystemClock implementations; that will come in a subsequent PR. It would be good to unify many of the Mock Timer implementations, so that they behave similarly and be tested similarly (some override Sleep, some use a MockSleep, etc). Additionally, this change will allow new methods to be introduced to the SystemClock (like https://github.com/facebook/rocksdb/issues/7101 WaitFor) in a consistent manner across a smaller number of classes. Pull Request resolved: https://github.com/facebook/rocksdb/pull/7858 Reviewed By: pdillinger Differential Revision: D26006406 Pulled By: mrambacher fbshipit-source-id: ed10a8abbdab7ff2e23d69d85bd25b3e7e899e90
511 lines
17 KiB
C++
511 lines
17 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
// Introduction of SyncPoint effectively disabled building and running this test
|
|
// in Release build.
|
|
// which is a pity, it is a good test
|
|
#if !defined(ROCKSDB_LITE)
|
|
|
|
#include "db/db_test_util.h"
|
|
#include "env/mock_env.h"
|
|
#include "port/port.h"
|
|
#include "port/stack_trace.h"
|
|
#include "util/random.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
class DBTestDynamicLevel : public DBTestBase {
|
|
public:
|
|
DBTestDynamicLevel()
|
|
: DBTestBase("/db_dynamic_level_test", /*env_do_fsync=*/true) {}
|
|
};
|
|
|
|
TEST_F(DBTestDynamicLevel, DynamicLevelMaxBytesBase) {
|
|
if (!Snappy_Supported() || !LZ4_Supported()) {
|
|
return;
|
|
}
|
|
// Use InMemoryEnv, or it would be too slow.
|
|
std::unique_ptr<Env> env(new MockEnv(env_));
|
|
|
|
const int kNKeys = 1000;
|
|
int keys[kNKeys];
|
|
|
|
auto verify_func = [&]() {
|
|
for (int i = 0; i < kNKeys; i++) {
|
|
ASSERT_NE("NOT_FOUND", Get(Key(i)));
|
|
ASSERT_NE("NOT_FOUND", Get(Key(kNKeys * 2 + i)));
|
|
if (i < kNKeys / 10) {
|
|
ASSERT_EQ("NOT_FOUND", Get(Key(kNKeys + keys[i])));
|
|
} else {
|
|
ASSERT_NE("NOT_FOUND", Get(Key(kNKeys + keys[i])));
|
|
}
|
|
}
|
|
};
|
|
|
|
Random rnd(301);
|
|
for (int ordered_insert = 0; ordered_insert <= 1; ordered_insert++) {
|
|
for (int i = 0; i < kNKeys; i++) {
|
|
keys[i] = i;
|
|
}
|
|
if (ordered_insert == 0) {
|
|
RandomShuffle(std::begin(keys), std::end(keys), rnd.Next());
|
|
}
|
|
for (int max_background_compactions = 1; max_background_compactions < 4;
|
|
max_background_compactions += 2) {
|
|
Options options;
|
|
options.env = env.get();
|
|
options.create_if_missing = true;
|
|
options.write_buffer_size = 2048;
|
|
options.max_write_buffer_number = 2;
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
options.level0_slowdown_writes_trigger = 2;
|
|
options.level0_stop_writes_trigger = 2;
|
|
options.target_file_size_base = 2048;
|
|
options.level_compaction_dynamic_level_bytes = true;
|
|
options.max_bytes_for_level_base = 10240;
|
|
options.max_bytes_for_level_multiplier = 4;
|
|
options.soft_rate_limit = 1.1;
|
|
options.max_background_compactions = max_background_compactions;
|
|
options.num_levels = 5;
|
|
|
|
options.compression_per_level.resize(3);
|
|
options.compression_per_level[0] = kNoCompression;
|
|
options.compression_per_level[1] = kLZ4Compression;
|
|
options.compression_per_level[2] = kSnappyCompression;
|
|
options.env = env_;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
for (int i = 0; i < kNKeys; i++) {
|
|
int key = keys[i];
|
|
ASSERT_OK(Put(Key(kNKeys + key), rnd.RandomString(102)));
|
|
ASSERT_OK(Put(Key(key), rnd.RandomString(102)));
|
|
ASSERT_OK(Put(Key(kNKeys * 2 + key), rnd.RandomString(102)));
|
|
ASSERT_OK(Delete(Key(kNKeys + keys[i / 10])));
|
|
env_->SleepForMicroseconds(5000);
|
|
}
|
|
|
|
uint64_t int_prop;
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.background-errors", &int_prop));
|
|
ASSERT_EQ(0U, int_prop);
|
|
|
|
// Verify DB
|
|
for (int j = 0; j < 2; j++) {
|
|
verify_func();
|
|
if (j == 0) {
|
|
Reopen(options);
|
|
}
|
|
}
|
|
|
|
// Test compact range works
|
|
ASSERT_OK(
|
|
dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
// All data should be in the last level.
|
|
ColumnFamilyMetaData cf_meta;
|
|
db_->GetColumnFamilyMetaData(&cf_meta);
|
|
ASSERT_EQ(5U, cf_meta.levels.size());
|
|
for (int i = 0; i < 4; i++) {
|
|
ASSERT_EQ(0U, cf_meta.levels[i].files.size());
|
|
}
|
|
ASSERT_GT(cf_meta.levels[4U].files.size(), 0U);
|
|
verify_func();
|
|
|
|
Close();
|
|
}
|
|
}
|
|
|
|
env_->SetBackgroundThreads(1, Env::LOW);
|
|
env_->SetBackgroundThreads(1, Env::HIGH);
|
|
}
|
|
|
|
// Test specific cases in dynamic max bytes
|
|
TEST_F(DBTestDynamicLevel, DynamicLevelMaxBytesBase2) {
|
|
Random rnd(301);
|
|
int kMaxKey = 1000000;
|
|
|
|
Options options = CurrentOptions();
|
|
options.compression = kNoCompression;
|
|
options.create_if_missing = true;
|
|
options.write_buffer_size = 20480;
|
|
options.max_write_buffer_number = 2;
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
options.level0_slowdown_writes_trigger = 9999;
|
|
options.level0_stop_writes_trigger = 9999;
|
|
options.target_file_size_base = 9102;
|
|
options.level_compaction_dynamic_level_bytes = true;
|
|
options.max_bytes_for_level_base = 40960;
|
|
options.max_bytes_for_level_multiplier = 4;
|
|
options.max_background_compactions = 2;
|
|
options.num_levels = 5;
|
|
options.max_compaction_bytes = 0; // Force not expanding in compactions
|
|
options.db_host_id = ""; // Setting this messes up the file size calculation
|
|
BlockBasedTableOptions table_options;
|
|
table_options.block_size = 1024;
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
DestroyAndReopen(options);
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "true"},
|
|
}));
|
|
|
|
uint64_t int_prop;
|
|
std::string str_prop;
|
|
|
|
// Initial base level is the last level
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(4U, int_prop);
|
|
|
|
// Put about 28K to L0
|
|
for (int i = 0; i < 70; i++) {
|
|
ASSERT_OK(Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))),
|
|
rnd.RandomString(380)));
|
|
}
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "false"},
|
|
}));
|
|
ASSERT_OK(Flush());
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(4U, int_prop);
|
|
|
|
// Insert extra about 28K to L0. After they are compacted to L4, the base
|
|
// level should be changed to L3.
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "true"},
|
|
}));
|
|
for (int i = 0; i < 70; i++) {
|
|
ASSERT_OK(Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))),
|
|
rnd.RandomString(380)));
|
|
}
|
|
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "false"},
|
|
}));
|
|
ASSERT_OK(Flush());
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(3U, int_prop);
|
|
ASSERT_TRUE(db_->GetProperty("rocksdb.num-files-at-level1", &str_prop));
|
|
ASSERT_EQ("0", str_prop);
|
|
ASSERT_TRUE(db_->GetProperty("rocksdb.num-files-at-level2", &str_prop));
|
|
ASSERT_EQ("0", str_prop);
|
|
|
|
// Write even more data while leaving the base level at L3.
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "true"},
|
|
}));
|
|
// Write about 40K more
|
|
for (int i = 0; i < 100; i++) {
|
|
ASSERT_OK(Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))),
|
|
rnd.RandomString(380)));
|
|
}
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "false"},
|
|
}));
|
|
ASSERT_OK(Flush());
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(3U, int_prop);
|
|
|
|
// Fill up L0, and then run an (auto) L0->Lmax compaction to raise the base
|
|
// level to 2.
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "true"},
|
|
}));
|
|
// Write about 650K more.
|
|
// Each file is about 11KB, with 9KB of data.
|
|
for (int i = 0; i < 1300; i++) {
|
|
ASSERT_OK(Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))),
|
|
rnd.RandomString(380)));
|
|
}
|
|
|
|
// Make sure that the compaction starts before the last bit of data is
|
|
// flushed, so that the base level isn't raised to L1.
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency({
|
|
{"CompactionJob::Run():Start", "DynamicLevelMaxBytesBase2:0"},
|
|
});
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "false"},
|
|
}));
|
|
|
|
TEST_SYNC_POINT("DynamicLevelMaxBytesBase2:0");
|
|
ASSERT_OK(Flush());
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(2U, int_prop);
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
// Write more data until the base level changes to L1. There will be
|
|
// a manual compaction going on at the same time.
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency({
|
|
{"CompactionJob::Run():Start", "DynamicLevelMaxBytesBase2:1"},
|
|
{"DynamicLevelMaxBytesBase2:2", "CompactionJob::Run():End"},
|
|
{"DynamicLevelMaxBytesBase2:compact_range_finish",
|
|
"FlushJob::WriteLevel0Table"},
|
|
});
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
ROCKSDB_NAMESPACE::port::Thread thread([this] {
|
|
TEST_SYNC_POINT("DynamicLevelMaxBytesBase2:compact_range_start");
|
|
ASSERT_OK(db_->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
TEST_SYNC_POINT("DynamicLevelMaxBytesBase2:compact_range_finish");
|
|
});
|
|
|
|
TEST_SYNC_POINT("DynamicLevelMaxBytesBase2:1");
|
|
for (int i = 0; i < 2; i++) {
|
|
ASSERT_OK(Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))),
|
|
rnd.RandomString(380)));
|
|
}
|
|
TEST_SYNC_POINT("DynamicLevelMaxBytesBase2:2");
|
|
|
|
ASSERT_OK(Flush());
|
|
|
|
thread.join();
|
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(1U, int_prop);
|
|
}
|
|
|
|
// Test specific cases in dynamic max bytes
|
|
TEST_F(DBTestDynamicLevel, DynamicLevelMaxBytesCompactRange) {
|
|
Random rnd(301);
|
|
int kMaxKey = 1000000;
|
|
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
options.write_buffer_size = 2048;
|
|
options.max_write_buffer_number = 2;
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
options.level0_slowdown_writes_trigger = 9999;
|
|
options.level0_stop_writes_trigger = 9999;
|
|
options.target_file_size_base = 2;
|
|
options.level_compaction_dynamic_level_bytes = true;
|
|
options.max_bytes_for_level_base = 10240;
|
|
options.max_bytes_for_level_multiplier = 4;
|
|
options.max_background_compactions = 1;
|
|
const int kNumLevels = 5;
|
|
options.num_levels = kNumLevels;
|
|
options.max_compaction_bytes = 1; // Force not expanding in compactions
|
|
BlockBasedTableOptions table_options;
|
|
table_options.block_size = 1024;
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
// Compact against empty DB
|
|
ASSERT_OK(dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
|
|
uint64_t int_prop;
|
|
std::string str_prop;
|
|
|
|
// Initial base level is the last level
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(4U, int_prop);
|
|
|
|
// Put about 7K to L0
|
|
for (int i = 0; i < 140; i++) {
|
|
ASSERT_OK(
|
|
Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))), rnd.RandomString(80)));
|
|
}
|
|
ASSERT_OK(Flush());
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
if (NumTableFilesAtLevel(0) == 0) {
|
|
// Make sure level 0 is not empty
|
|
ASSERT_OK(
|
|
Put(Key(static_cast<int>(rnd.Uniform(kMaxKey))), rnd.RandomString(80)));
|
|
ASSERT_OK(Flush());
|
|
}
|
|
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(3U, int_prop);
|
|
ASSERT_TRUE(db_->GetProperty("rocksdb.num-files-at-level1", &str_prop));
|
|
ASSERT_EQ("0", str_prop);
|
|
ASSERT_TRUE(db_->GetProperty("rocksdb.num-files-at-level2", &str_prop));
|
|
ASSERT_EQ("0", str_prop);
|
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
std::set<int> output_levels;
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
|
|
"CompactionPicker::CompactRange:Return", [&](void* arg) {
|
|
Compaction* compaction = reinterpret_cast<Compaction*>(arg);
|
|
output_levels.insert(compaction->output_level());
|
|
});
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
ASSERT_OK(dbfull()->CompactRange(CompactRangeOptions(), nullptr, nullptr));
|
|
ASSERT_EQ(output_levels.size(), 2);
|
|
ASSERT_TRUE(output_levels.find(3) != output_levels.end());
|
|
ASSERT_TRUE(output_levels.find(4) != output_levels.end());
|
|
ASSERT_TRUE(db_->GetProperty("rocksdb.num-files-at-level0", &str_prop));
|
|
ASSERT_EQ("0", str_prop);
|
|
ASSERT_TRUE(db_->GetProperty("rocksdb.num-files-at-level3", &str_prop));
|
|
ASSERT_EQ("0", str_prop);
|
|
// Base level is still level 3.
|
|
ASSERT_TRUE(db_->GetIntProperty("rocksdb.base-level", &int_prop));
|
|
ASSERT_EQ(3U, int_prop);
|
|
}
|
|
|
|
TEST_F(DBTestDynamicLevel, DynamicLevelMaxBytesBaseInc) {
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
options.write_buffer_size = 2048;
|
|
options.max_write_buffer_number = 2;
|
|
options.level0_file_num_compaction_trigger = 2;
|
|
options.level0_slowdown_writes_trigger = 2;
|
|
options.level0_stop_writes_trigger = 2;
|
|
options.target_file_size_base = 2048;
|
|
options.level_compaction_dynamic_level_bytes = true;
|
|
options.max_bytes_for_level_base = 10240;
|
|
options.max_bytes_for_level_multiplier = 4;
|
|
options.soft_rate_limit = 1.1;
|
|
options.max_background_compactions = 2;
|
|
options.num_levels = 5;
|
|
options.max_compaction_bytes = 100000000;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
int non_trivial = 0;
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->SetCallBack(
|
|
"DBImpl::BackgroundCompaction:NonTrivial",
|
|
[&](void* /*arg*/) { non_trivial++; });
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
Random rnd(301);
|
|
const int total_keys = 3000;
|
|
const int random_part_size = 100;
|
|
for (int i = 0; i < total_keys; i++) {
|
|
std::string value = rnd.RandomString(random_part_size);
|
|
PutFixed32(&value, static_cast<uint32_t>(i));
|
|
ASSERT_OK(Put(Key(i), value));
|
|
}
|
|
ASSERT_OK(Flush());
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
ASSERT_EQ(non_trivial, 0);
|
|
|
|
for (int i = 0; i < total_keys; i++) {
|
|
std::string value = Get(Key(i));
|
|
ASSERT_EQ(DecodeFixed32(value.c_str() + random_part_size),
|
|
static_cast<uint32_t>(i));
|
|
}
|
|
|
|
env_->SetBackgroundThreads(1, Env::LOW);
|
|
env_->SetBackgroundThreads(1, Env::HIGH);
|
|
}
|
|
|
|
TEST_F(DBTestDynamicLevel, DISABLED_MigrateToDynamicLevelMaxBytesBase) {
|
|
Random rnd(301);
|
|
const int kMaxKey = 2000;
|
|
|
|
Options options;
|
|
options.create_if_missing = true;
|
|
options.write_buffer_size = 2048;
|
|
options.max_write_buffer_number = 8;
|
|
options.level0_file_num_compaction_trigger = 4;
|
|
options.level0_slowdown_writes_trigger = 4;
|
|
options.level0_stop_writes_trigger = 8;
|
|
options.target_file_size_base = 2048;
|
|
options.level_compaction_dynamic_level_bytes = false;
|
|
options.max_bytes_for_level_base = 10240;
|
|
options.max_bytes_for_level_multiplier = 4;
|
|
options.soft_rate_limit = 1.1;
|
|
options.num_levels = 8;
|
|
|
|
DestroyAndReopen(options);
|
|
|
|
auto verify_func = [&](int num_keys, bool if_sleep) {
|
|
for (int i = 0; i < num_keys; i++) {
|
|
ASSERT_NE("NOT_FOUND", Get(Key(kMaxKey + i)));
|
|
if (i < num_keys / 10) {
|
|
ASSERT_EQ("NOT_FOUND", Get(Key(i)));
|
|
} else {
|
|
ASSERT_NE("NOT_FOUND", Get(Key(i)));
|
|
}
|
|
if (if_sleep && i % 1000 == 0) {
|
|
// Without it, valgrind may choose not to give another
|
|
// thread a chance to run before finishing the function,
|
|
// causing the test to be extremely slow.
|
|
env_->SleepForMicroseconds(1);
|
|
}
|
|
}
|
|
};
|
|
|
|
int total_keys = 1000;
|
|
for (int i = 0; i < total_keys; i++) {
|
|
ASSERT_OK(Put(Key(i), rnd.RandomString(102)));
|
|
ASSERT_OK(Put(Key(kMaxKey + i), rnd.RandomString(102)));
|
|
ASSERT_OK(Delete(Key(i / 10)));
|
|
}
|
|
verify_func(total_keys, false);
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
|
|
options.level_compaction_dynamic_level_bytes = true;
|
|
options.disable_auto_compactions = true;
|
|
Reopen(options);
|
|
verify_func(total_keys, false);
|
|
|
|
std::atomic_bool compaction_finished;
|
|
compaction_finished = false;
|
|
// Issue manual compaction in one thread and still verify DB state
|
|
// in main thread.
|
|
ROCKSDB_NAMESPACE::port::Thread t([&]() {
|
|
CompactRangeOptions compact_options;
|
|
compact_options.change_level = true;
|
|
compact_options.target_level = options.num_levels - 1;
|
|
ASSERT_OK(dbfull()->CompactRange(compact_options, nullptr, nullptr));
|
|
compaction_finished.store(true);
|
|
});
|
|
do {
|
|
verify_func(total_keys, true);
|
|
} while (!compaction_finished.load());
|
|
t.join();
|
|
|
|
ASSERT_OK(dbfull()->SetOptions({
|
|
{"disable_auto_compactions", "false"},
|
|
}));
|
|
|
|
int total_keys2 = 2000;
|
|
for (int i = total_keys; i < total_keys2; i++) {
|
|
ASSERT_OK(Put(Key(i), rnd.RandomString(102)));
|
|
ASSERT_OK(Put(Key(kMaxKey + i), rnd.RandomString(102)));
|
|
ASSERT_OK(Delete(Key(i / 10)));
|
|
}
|
|
|
|
verify_func(total_keys2, false);
|
|
ASSERT_OK(dbfull()->TEST_WaitForCompact());
|
|
verify_func(total_keys2, false);
|
|
|
|
// Base level is not level 1
|
|
ASSERT_EQ(NumTableFilesAtLevel(1), 0);
|
|
ASSERT_EQ(NumTableFilesAtLevel(2), 0);
|
|
}
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
#endif // !defined(ROCKSDB_LITE)
|
|
|
|
int main(int argc, char** argv) {
|
|
#if !defined(ROCKSDB_LITE)
|
|
ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
return RUN_ALL_TESTS();
|
|
#else
|
|
(void) argc;
|
|
(void) argv;
|
|
return 0;
|
|
#endif
|
|
}
|