2016-02-09 23:12:00 +00:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-15 23:03:42 +00:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2013-10-16 21:59:46 +00:00
|
|
|
//
|
2013-08-22 21:32:53 +00:00
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
2015-07-20 18:24:54 +00:00
|
|
|
|
2016-06-22 01:41:23 +00:00
|
|
|
#include <stdlib.h>
|
2022-11-02 21:34:24 +00:00
|
|
|
|
2016-06-22 01:41:23 +00:00
|
|
|
#include <map>
|
|
|
|
#include <string>
|
|
|
|
#include <vector>
|
2022-11-02 21:34:24 +00:00
|
|
|
|
2019-05-31 18:52:59 +00:00
|
|
|
#include "db/db_impl/db_impl.h"
|
2019-09-18 23:56:25 +00:00
|
|
|
#include "db/db_test_util.h"
|
2013-08-22 21:32:53 +00:00
|
|
|
#include "db/version_set.h"
|
|
|
|
#include "db/write_batch_internal.h"
|
2019-05-30 03:44:08 +00:00
|
|
|
#include "file/filename.h"
|
2019-09-18 23:56:25 +00:00
|
|
|
#include "port/stack_trace.h"
|
2016-06-22 01:41:23 +00:00
|
|
|
#include "rocksdb/db.h"
|
|
|
|
#include "rocksdb/env.h"
|
|
|
|
#include "rocksdb/transaction_log.h"
|
2019-05-30 18:21:38 +00:00
|
|
|
#include "test_util/sync_point.h"
|
|
|
|
#include "test_util/testharness.h"
|
|
|
|
#include "test_util/testutil.h"
|
2019-05-31 00:39:43 +00:00
|
|
|
#include "util/string_util.h"
|
2013-08-22 21:32:53 +00:00
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
2013-08-22 21:32:53 +00:00
|
|
|
|
2019-09-18 23:56:25 +00:00
|
|
|
class DeleteFileTest : public DBTestBase {
|
2013-08-22 21:32:53 +00:00
|
|
|
public:
|
2019-09-18 23:56:25 +00:00
|
|
|
const int numlevels_;
|
|
|
|
const std::string wal_dir_;
|
|
|
|
|
|
|
|
DeleteFileTest()
|
2021-07-23 15:37:27 +00:00
|
|
|
: DBTestBase("deletefile_test", /*env_do_fsync=*/true),
|
2019-09-18 23:56:25 +00:00
|
|
|
numlevels_(7),
|
|
|
|
wal_dir_(dbname_ + "/wal_files") {}
|
|
|
|
|
|
|
|
void SetOptions(Options* options) {
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_NE(options, nullptr);
|
2019-09-18 23:56:25 +00:00
|
|
|
options->delete_obsolete_files_period_micros = 0; // always do full purge
|
|
|
|
options->enable_thread_tracking = true;
|
|
|
|
options->write_buffer_size = 1024 * 1024 * 1000;
|
|
|
|
options->target_file_size_base = 1024 * 1024 * 1000;
|
|
|
|
options->max_bytes_for_level_base = 1024 * 1024 * 1000;
|
|
|
|
options->WAL_ttl_seconds = 300; // Used to test log files
|
|
|
|
options->WAL_size_limit_MB = 1024; // Used to test log files
|
|
|
|
options->wal_dir = wal_dir_;
|
2013-08-22 21:32:53 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void AddKeys(int numkeys, int startkey = 0) {
|
|
|
|
WriteOptions options;
|
|
|
|
options.sync = false;
|
|
|
|
ReadOptions roptions;
|
2022-11-02 21:34:24 +00:00
|
|
|
for (int i = startkey; i < (numkeys + startkey); i++) {
|
2022-05-06 20:03:58 +00:00
|
|
|
std::string temp = std::to_string(i);
|
2013-08-22 21:32:53 +00:00
|
|
|
Slice key(temp);
|
|
|
|
Slice value(temp);
|
|
|
|
ASSERT_OK(db_->Put(options, key, value));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-11-02 21:34:24 +00:00
|
|
|
int numKeysInLevels(std::vector<LiveFileMetaData>& metadata,
|
|
|
|
std::vector<int>* keysperlevel = nullptr) {
|
2013-08-22 21:32:53 +00:00
|
|
|
if (keysperlevel != nullptr) {
|
|
|
|
keysperlevel->resize(numlevels_);
|
|
|
|
}
|
|
|
|
|
|
|
|
int numKeys = 0;
|
|
|
|
for (size_t i = 0; i < metadata.size(); i++) {
|
|
|
|
int startkey = atoi(metadata[i].smallestkey.c_str());
|
|
|
|
int endkey = atoi(metadata[i].largestkey.c_str());
|
|
|
|
int numkeysinfile = (endkey - startkey + 1);
|
|
|
|
numKeys += numkeysinfile;
|
|
|
|
if (keysperlevel != nullptr) {
|
|
|
|
(*keysperlevel)[(int)metadata[i].level] += numkeysinfile;
|
|
|
|
}
|
|
|
|
fprintf(stderr, "level %d name %s smallest %s largest %s\n",
|
|
|
|
metadata[i].level, metadata[i].name.c_str(),
|
2022-11-02 21:34:24 +00:00
|
|
|
metadata[i].smallestkey.c_str(), metadata[i].largestkey.c_str());
|
2013-08-22 21:32:53 +00:00
|
|
|
}
|
|
|
|
return numKeys;
|
|
|
|
}
|
|
|
|
|
|
|
|
void CreateTwoLevels() {
|
|
|
|
AddKeys(50000, 10000);
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(dbfull()->TEST_FlushMemTable());
|
|
|
|
ASSERT_OK(dbfull()->TEST_WaitForFlushMemTable());
|
2015-07-17 19:02:52 +00:00
|
|
|
for (int i = 0; i < 2; ++i) {
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(dbfull()->TEST_CompactRange(i, nullptr, nullptr));
|
2015-07-17 19:02:52 +00:00
|
|
|
}
|
2013-08-22 21:32:53 +00:00
|
|
|
|
|
|
|
AddKeys(50000, 10000);
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(dbfull()->TEST_FlushMemTable());
|
|
|
|
ASSERT_OK(dbfull()->TEST_WaitForFlushMemTable());
|
|
|
|
ASSERT_OK(dbfull()->TEST_CompactRange(0, nullptr, nullptr));
|
2013-08-22 21:32:53 +00:00
|
|
|
}
|
|
|
|
|
2019-09-18 23:56:25 +00:00
|
|
|
void CheckFileTypeCounts(const std::string& dir, int required_log,
|
|
|
|
int required_sst, int required_manifest) {
|
2013-11-15 02:03:57 +00:00
|
|
|
std::vector<std::string> filenames;
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(env_->GetChildren(dir, &filenames));
|
2013-11-15 02:03:57 +00:00
|
|
|
|
|
|
|
int log_cnt = 0, sst_cnt = 0, manifest_cnt = 0;
|
|
|
|
for (auto file : filenames) {
|
|
|
|
uint64_t number;
|
|
|
|
FileType type;
|
|
|
|
if (ParseFileName(file, &number, &type)) {
|
2020-10-23 00:04:39 +00:00
|
|
|
log_cnt += (type == kWalFile);
|
2013-11-15 02:03:57 +00:00
|
|
|
sst_cnt += (type == kTableFile);
|
|
|
|
manifest_cnt += (type == kDescriptorFile);
|
|
|
|
}
|
|
|
|
}
|
2021-11-03 19:20:19 +00:00
|
|
|
if (required_log >= 0) {
|
|
|
|
ASSERT_EQ(required_log, log_cnt);
|
|
|
|
}
|
|
|
|
if (required_sst >= 0) {
|
|
|
|
ASSERT_EQ(required_sst, sst_cnt);
|
|
|
|
}
|
|
|
|
if (required_manifest >= 0) {
|
|
|
|
ASSERT_EQ(required_manifest, manifest_cnt);
|
|
|
|
}
|
2013-11-15 02:03:57 +00:00
|
|
|
}
|
|
|
|
|
2016-06-22 01:41:23 +00:00
|
|
|
static void DoSleep(void* arg) {
|
|
|
|
auto test = reinterpret_cast<DeleteFileTest*>(arg);
|
|
|
|
test->env_->SleepForMicroseconds(2 * 1000 * 1000);
|
|
|
|
}
|
|
|
|
|
|
|
|
// An empty job to guard all jobs are processed
|
2018-03-05 21:08:17 +00:00
|
|
|
static void GuardFinish(void* /*arg*/) {
|
2016-06-22 01:41:23 +00:00
|
|
|
TEST_SYNC_POINT("DeleteFileTest::GuardFinish");
|
|
|
|
}
|
2013-08-22 21:32:53 +00:00
|
|
|
};
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(DeleteFileTest, AddKeysAndQueryLevels) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2013-08-22 21:32:53 +00:00
|
|
|
CreateTwoLevels();
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
|
|
|
|
std::string level1file = "";
|
|
|
|
int level1keycount = 0;
|
|
|
|
std::string level2file = "";
|
|
|
|
int level2keycount = 0;
|
|
|
|
int level1index = 0;
|
|
|
|
int level2index = 1;
|
|
|
|
|
|
|
|
ASSERT_EQ((int)metadata.size(), 2);
|
|
|
|
if (metadata[0].level == 2) {
|
|
|
|
level1index = 1;
|
|
|
|
level2index = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
level1file = metadata[level1index].name;
|
|
|
|
int startkey = atoi(metadata[level1index].smallestkey.c_str());
|
|
|
|
int endkey = atoi(metadata[level1index].largestkey.c_str());
|
|
|
|
level1keycount = (endkey - startkey + 1);
|
|
|
|
level2file = metadata[level2index].name;
|
|
|
|
startkey = atoi(metadata[level2index].smallestkey.c_str());
|
|
|
|
endkey = atoi(metadata[level2index].largestkey.c_str());
|
|
|
|
level2keycount = (endkey - startkey + 1);
|
|
|
|
|
|
|
|
// COntrolled setup. Levels 1 and 2 should both have 50K files.
|
|
|
|
// This is a little fragile as it depends on the current
|
|
|
|
// compaction heuristics.
|
|
|
|
ASSERT_EQ(level1keycount, 50000);
|
|
|
|
ASSERT_EQ(level2keycount, 50000);
|
|
|
|
|
|
|
|
Status status = db_->DeleteFile("0.sst");
|
|
|
|
ASSERT_TRUE(status.IsInvalidArgument());
|
|
|
|
|
|
|
|
// intermediate level files cannot be deleted.
|
|
|
|
status = db_->DeleteFile(level1file);
|
|
|
|
ASSERT_TRUE(status.IsInvalidArgument());
|
|
|
|
|
|
|
|
// Lowest level file deletion should succeed.
|
2020-12-22 23:08:17 +00:00
|
|
|
status = db_->DeleteFile(level2file);
|
|
|
|
ASSERT_OK(status);
|
2013-08-22 21:32:53 +00:00
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(DeleteFileTest, PurgeObsoleteFilesTest) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2013-11-15 02:03:57 +00:00
|
|
|
CreateTwoLevels();
|
|
|
|
// there should be only one (empty) log file because CreateTwoLevels()
|
|
|
|
// flushes the memtables to disk
|
2019-09-18 23:56:25 +00:00
|
|
|
CheckFileTypeCounts(wal_dir_, 1, 0, 0);
|
2013-11-15 02:03:57 +00:00
|
|
|
// 2 ssts, 1 manifest
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 2, 1);
|
|
|
|
std::string first("0"), last("999999");
|
2015-06-17 21:36:14 +00:00
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
2013-11-15 02:03:57 +00:00
|
|
|
Slice first_slice(first), last_slice(last);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(db_->CompactRange(compact_options, &first_slice, &last_slice));
|
2013-11-15 02:03:57 +00:00
|
|
|
// 1 sst after compaction
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
// this time, we keep an iterator alive
|
2019-09-18 23:56:25 +00:00
|
|
|
Reopen(options);
|
2022-11-02 21:34:24 +00:00
|
|
|
Iterator* itr = nullptr;
|
2013-11-15 02:03:57 +00:00
|
|
|
CreateTwoLevels();
|
|
|
|
itr = db_->NewIterator(ReadOptions());
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(itr->status());
|
|
|
|
ASSERT_OK(db_->CompactRange(compact_options, &first_slice, &last_slice));
|
|
|
|
ASSERT_OK(itr->status());
|
2013-11-15 02:03:57 +00:00
|
|
|
// 3 sst after compaction with live iterator
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
delete itr;
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
}
|
|
|
|
|
2019-04-02 00:07:38 +00:00
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeIteratorTest) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2016-06-22 01:41:23 +00:00
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
|
|
|
|
// We keep an iterator alive
|
2018-03-07 20:39:19 +00:00
|
|
|
Iterator* itr = nullptr;
|
2016-06-22 01:41:23 +00:00
|
|
|
CreateTwoLevels();
|
2019-09-18 23:56:25 +00:00
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.background_purge_on_iterator_cleanup = true;
|
|
|
|
itr = db_->NewIterator(read_options);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(itr->status());
|
|
|
|
ASSERT_OK(db_->CompactRange(compact_options, &first_slice, &last_slice));
|
2016-06-22 01:41:23 +00:00
|
|
|
// 3 sst after compaction with live iterator
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
test::SleepingBackgroundTask sleeping_task_before;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_before, Env::Priority::HIGH);
|
|
|
|
delete itr;
|
|
|
|
test::SleepingBackgroundTask sleeping_task_after;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_after, Env::Priority::HIGH);
|
|
|
|
|
|
|
|
// Make sure no purges are executed foreground
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
sleeping_task_before.WakeUp();
|
|
|
|
sleeping_task_before.WaitUntilDone();
|
|
|
|
|
|
|
|
// Make sure all background purges are executed
|
|
|
|
sleeping_task_after.WakeUp();
|
|
|
|
sleeping_task_after.WaitUntilDone();
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
}
|
|
|
|
|
2021-11-03 19:20:19 +00:00
|
|
|
TEST_F(DeleteFileTest, PurgeDuringOpen) {
|
|
|
|
Options options = CurrentOptions();
|
|
|
|
CheckFileTypeCounts(dbname_, -1, 0, -1);
|
|
|
|
Close();
|
|
|
|
std::unique_ptr<WritableFile> file;
|
|
|
|
ASSERT_OK(options.env->NewWritableFile(dbname_ + "/000002.sst", &file,
|
|
|
|
EnvOptions()));
|
|
|
|
ASSERT_OK(file->Close());
|
|
|
|
CheckFileTypeCounts(dbname_, -1, 1, -1);
|
|
|
|
options.avoid_unnecessary_blocking_io = false;
|
|
|
|
options.create_if_missing = false;
|
|
|
|
Reopen(options);
|
|
|
|
CheckFileTypeCounts(dbname_, -1, 0, -1);
|
|
|
|
Close();
|
|
|
|
|
|
|
|
// test background purge
|
|
|
|
options.avoid_unnecessary_blocking_io = true;
|
|
|
|
options.create_if_missing = false;
|
|
|
|
ASSERT_OK(options.env->NewWritableFile(dbname_ + "/000002.sst", &file,
|
|
|
|
EnvOptions()));
|
|
|
|
ASSERT_OK(file->Close());
|
|
|
|
CheckFileTypeCounts(dbname_, -1, 1, -1);
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency(
|
|
|
|
{{"DeleteFileTest::PurgeDuringOpen:1", "DBImpl::BGWorkPurge:start"}});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
Reopen(options);
|
|
|
|
// the obsolete file is not deleted until the background purge job is ran
|
|
|
|
CheckFileTypeCounts(dbname_, -1, 1, -1);
|
|
|
|
TEST_SYNC_POINT("DeleteFileTest::PurgeDuringOpen:1");
|
|
|
|
ASSERT_OK(dbfull()->TEST_WaitForPurge());
|
|
|
|
CheckFileTypeCounts(dbname_, -1, 0, -1);
|
|
|
|
}
|
|
|
|
|
2019-04-02 00:07:38 +00:00
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeCFDropTest) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2019-04-02 00:07:38 +00:00
|
|
|
auto do_test = [&](bool bg_purge) {
|
|
|
|
ColumnFamilyOptions co;
|
Refactor trimming logic for immutable memtables (#5022)
Summary:
MyRocks currently sets `max_write_buffer_number_to_maintain` in order to maintain enough history for transaction conflict checking. The effectiveness of this approach depends on the size of memtables. When memtables are small, it may not keep enough history; when memtables are large, this may consume too much memory.
We are proposing a new way to configure memtable list history: by limiting the memory usage of immutable memtables. The new option is `max_write_buffer_size_to_maintain` and it will take precedence over the old `max_write_buffer_number_to_maintain` if they are both set to non-zero values. The new option accounts for the total memory usage of flushed immutable memtables and mutable memtable. When the total usage exceeds the limit, RocksDB may start dropping immutable memtables (which is also called trimming history), starting from the oldest one.
The semantics of the old option actually works both as an upper bound and lower bound. History trimming will start if number of immutable memtables exceeds the limit, but it will never go below (limit-1) due to history trimming.
In order the mimic the behavior with the new option, history trimming will stop if dropping the next immutable memtable causes the total memory usage go below the size limit. For example, assuming the size limit is set to 64MB, and there are 3 immutable memtables with sizes of 20, 30, 30. Although the total memory usage is 80MB > 64MB, dropping the oldest memtable will reduce the memory usage to 60MB < 64MB, so in this case no memtable will be dropped.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5022
Differential Revision: D14394062
Pulled By: miasantreble
fbshipit-source-id: 60457a509c6af89d0993f988c9b5c2aa9e45f5c5
2019-08-23 20:54:09 +00:00
|
|
|
co.max_write_buffer_size_to_maintain =
|
|
|
|
static_cast<int64_t>(co.write_buffer_size);
|
2019-04-02 00:07:38 +00:00
|
|
|
WriteOptions wo;
|
|
|
|
FlushOptions fo;
|
|
|
|
ColumnFamilyHandle* cfh = nullptr;
|
|
|
|
|
|
|
|
ASSERT_OK(db_->CreateColumnFamily(co, "dropme", &cfh));
|
|
|
|
|
|
|
|
ASSERT_OK(db_->Put(wo, cfh, "pika", "chu"));
|
|
|
|
ASSERT_OK(db_->Flush(fo, cfh));
|
|
|
|
// Expect 1 sst file.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
ASSERT_OK(db_->DropColumnFamily(cfh));
|
|
|
|
// Still 1 file, it won't be deleted while ColumnFamilyHandle is alive.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
|
|
|
|
delete cfh;
|
|
|
|
test::SleepingBackgroundTask sleeping_task_after;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_after, Env::Priority::HIGH);
|
|
|
|
// If background purge is enabled, the file should still be there.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, bg_purge ? 1 : 0, 1);
|
2019-05-15 22:13:44 +00:00
|
|
|
TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeCFDropTest:1");
|
2019-04-02 00:07:38 +00:00
|
|
|
|
|
|
|
// Execute background purges.
|
|
|
|
sleeping_task_after.WakeUp();
|
|
|
|
sleeping_task_after.WaitUntilDone();
|
|
|
|
// The file should have been deleted.
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 0, 1);
|
|
|
|
};
|
|
|
|
|
|
|
|
{
|
|
|
|
SCOPED_TRACE("avoid_unnecessary_blocking_io = false");
|
|
|
|
do_test(false);
|
|
|
|
}
|
|
|
|
|
2021-11-03 19:20:19 +00:00
|
|
|
options.avoid_unnecessary_blocking_io = true;
|
|
|
|
options.create_if_missing = false;
|
|
|
|
Reopen(options);
|
|
|
|
ASSERT_OK(dbfull()->TEST_WaitForPurge());
|
|
|
|
|
2019-05-15 22:13:44 +00:00
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
|
|
SyncPoint::GetInstance()->LoadDependency(
|
|
|
|
{{"DeleteFileTest::BackgroundPurgeCFDropTest:1",
|
|
|
|
"DBImpl::BGWorkPurge:start"}});
|
|
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
|
2019-04-02 00:07:38 +00:00
|
|
|
{
|
|
|
|
SCOPED_TRACE("avoid_unnecessary_blocking_io = true");
|
|
|
|
do_test(true);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-07-05 18:57:14 +00:00
|
|
|
// This test is to reproduce a bug that read invalid ReadOption in iterator
|
|
|
|
// cleanup function
|
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeCopyOptions) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2016-07-05 18:57:14 +00:00
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
|
|
|
|
// We keep an iterator alive
|
2018-03-07 20:39:19 +00:00
|
|
|
Iterator* itr = nullptr;
|
2016-07-05 18:57:14 +00:00
|
|
|
CreateTwoLevels();
|
2019-09-18 23:56:25 +00:00
|
|
|
{
|
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.background_purge_on_iterator_cleanup = true;
|
|
|
|
itr = db_->NewIterator(read_options);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(itr->status());
|
2019-09-18 23:56:25 +00:00
|
|
|
// ReadOptions is deleted, but iterator cleanup function should not be
|
|
|
|
// affected
|
|
|
|
}
|
2016-07-05 18:57:14 +00:00
|
|
|
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(db_->CompactRange(compact_options, &first_slice, &last_slice));
|
2016-07-05 18:57:14 +00:00
|
|
|
// 3 sst after compaction with live iterator
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
|
|
delete itr;
|
|
|
|
|
|
|
|
test::SleepingBackgroundTask sleeping_task_after;
|
|
|
|
env_->Schedule(&test::SleepingBackgroundTask::DoSleepTask,
|
|
|
|
&sleeping_task_after, Env::Priority::HIGH);
|
|
|
|
|
|
|
|
// Make sure all background purges are executed
|
|
|
|
sleeping_task_after.WakeUp();
|
|
|
|
sleeping_task_after.WaitUntilDone();
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
}
|
|
|
|
|
2016-06-22 01:41:23 +00:00
|
|
|
TEST_F(DeleteFileTest, BackgroundPurgeTestMultipleJobs) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2016-06-22 01:41:23 +00:00
|
|
|
std::string first("0"), last("999999");
|
|
|
|
CompactRangeOptions compact_options;
|
|
|
|
compact_options.change_level = true;
|
|
|
|
compact_options.target_level = 2;
|
|
|
|
Slice first_slice(first), last_slice(last);
|
|
|
|
|
|
|
|
// We keep an iterator alive
|
|
|
|
CreateTwoLevels();
|
2019-09-18 23:56:25 +00:00
|
|
|
ReadOptions read_options;
|
|
|
|
read_options.background_purge_on_iterator_cleanup = true;
|
|
|
|
Iterator* itr1 = db_->NewIterator(read_options);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(itr1->status());
|
2016-06-22 01:41:23 +00:00
|
|
|
CreateTwoLevels();
|
2019-09-18 23:56:25 +00:00
|
|
|
Iterator* itr2 = db_->NewIterator(read_options);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(itr2->status());
|
|
|
|
ASSERT_OK(db_->CompactRange(compact_options, &first_slice, &last_slice));
|
2016-06-22 01:41:23 +00:00
|
|
|
// 5 sst files after 2 compactions with 2 live iterators
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 5, 1);
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->DisableProcessing();
|
2016-06-22 01:41:23 +00:00
|
|
|
// ~DBImpl should wait until all BGWorkPurge are finished
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->LoadDependency(
|
2016-06-22 01:41:23 +00:00
|
|
|
{{"DBImpl::~DBImpl:WaitJob", "DBImpl::BGWorkPurge"},
|
|
|
|
{"DeleteFileTest::GuardFinish",
|
|
|
|
"DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose"}});
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::SyncPoint::GetInstance()->EnableProcessing();
|
2016-06-22 01:41:23 +00:00
|
|
|
|
|
|
|
delete itr1;
|
|
|
|
env_->Schedule(&DeleteFileTest::DoSleep, this, Env::Priority::HIGH);
|
|
|
|
delete itr2;
|
|
|
|
env_->Schedule(&DeleteFileTest::GuardFinish, nullptr, Env::Priority::HIGH);
|
2019-09-18 23:56:25 +00:00
|
|
|
Close();
|
2016-06-22 01:41:23 +00:00
|
|
|
|
|
|
|
TEST_SYNC_POINT("DeleteFileTest::BackgroundPurgeTestMultipleJobs:DBClose");
|
|
|
|
// 1 sst after iterator deletion
|
|
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(DeleteFileTest, DeleteFileWithIterator) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2013-08-22 21:32:53 +00:00
|
|
|
CreateTwoLevels();
|
2019-09-18 23:56:25 +00:00
|
|
|
ReadOptions read_options;
|
|
|
|
Iterator* it = db_->NewIterator(read_options);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(it->status());
|
2013-08-22 21:32:53 +00:00
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
|
2019-09-18 23:56:25 +00:00
|
|
|
std::string level2file;
|
2013-08-22 21:32:53 +00:00
|
|
|
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_EQ(metadata.size(), static_cast<size_t>(2));
|
2013-08-22 21:32:53 +00:00
|
|
|
if (metadata[0].level == 1) {
|
|
|
|
level2file = metadata[1].name;
|
|
|
|
} else {
|
|
|
|
level2file = metadata[0].name;
|
|
|
|
}
|
|
|
|
|
|
|
|
Status status = db_->DeleteFile(level2file);
|
2022-11-02 21:34:24 +00:00
|
|
|
fprintf(stdout, "Deletion status %s: %s\n", level2file.c_str(),
|
|
|
|
status.ToString().c_str());
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(status);
|
2013-08-22 21:32:53 +00:00
|
|
|
it->SeekToFirst();
|
|
|
|
int numKeysIterated = 0;
|
2022-11-02 21:34:24 +00:00
|
|
|
while (it->Valid()) {
|
2013-08-22 21:32:53 +00:00
|
|
|
numKeysIterated++;
|
|
|
|
it->Next();
|
|
|
|
}
|
|
|
|
ASSERT_EQ(numKeysIterated, 50000);
|
|
|
|
delete it;
|
|
|
|
}
|
2013-10-24 06:39:23 +00:00
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(DeleteFileTest, DeleteLogFiles) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
|
2013-10-24 06:39:23 +00:00
|
|
|
AddKeys(10, 0);
|
|
|
|
VectorLogPtr logfiles;
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(db_->GetSortedWalFiles(logfiles));
|
2013-10-24 06:39:23 +00:00
|
|
|
ASSERT_GT(logfiles.size(), 0UL);
|
|
|
|
// Take the last log file which is expected to be alive and try to delete it
|
|
|
|
// Should not succeed because live logs are not allowed to be deleted
|
|
|
|
std::unique_ptr<LogFile> alive_log = std::move(logfiles.back());
|
|
|
|
ASSERT_EQ(alive_log->Type(), kAliveLogFile);
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(env_->FileExists(wal_dir_ + "/" + alive_log->PathName()));
|
2013-10-24 06:39:23 +00:00
|
|
|
fprintf(stdout, "Deleting alive log file %s\n",
|
|
|
|
alive_log->PathName().c_str());
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_NOK(db_->DeleteFile(alive_log->PathName()));
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(env_->FileExists(wal_dir_ + "/" + alive_log->PathName()));
|
2013-10-24 06:39:23 +00:00
|
|
|
logfiles.clear();
|
|
|
|
|
|
|
|
// Call Flush to bring about a new working log file and add more keys
|
|
|
|
// Call Flush again to flush out memtable and move alive log to archived log
|
|
|
|
// and try to delete the archived log file
|
|
|
|
FlushOptions fopts;
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(db_->Flush(fopts));
|
2013-10-24 06:39:23 +00:00
|
|
|
AddKeys(10, 0);
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(db_->Flush(fopts));
|
|
|
|
ASSERT_OK(db_->GetSortedWalFiles(logfiles));
|
2013-10-24 06:39:23 +00:00
|
|
|
ASSERT_GT(logfiles.size(), 0UL);
|
|
|
|
std::unique_ptr<LogFile> archived_log = std::move(logfiles.front());
|
|
|
|
ASSERT_EQ(archived_log->Type(), kArchivedLogFile);
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(env_->FileExists(wal_dir_ + "/" + archived_log->PathName()));
|
2013-10-24 06:39:23 +00:00
|
|
|
fprintf(stdout, "Deleting archived log file %s\n",
|
|
|
|
archived_log->PathName().c_str());
|
|
|
|
ASSERT_OK(db_->DeleteFile(archived_log->PathName()));
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_TRUE(
|
|
|
|
env_->FileExists(wal_dir_ + "/" + archived_log->PathName()).IsNotFound());
|
2013-10-24 06:39:23 +00:00
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(DeleteFileTest, DeleteNonDefaultColumnFamily) {
|
2019-09-18 23:56:25 +00:00
|
|
|
Options options = CurrentOptions();
|
|
|
|
SetOptions(&options);
|
|
|
|
Destroy(options);
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(options);
|
|
|
|
CreateAndReopenWithCF({"new_cf"}, options);
|
2014-10-21 18:23:06 +00:00
|
|
|
|
|
|
|
Random rnd(5);
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), handles_[1], test::RandomKey(&rnd, 10),
|
|
|
|
test::RandomKey(&rnd, 10)));
|
2014-10-21 18:23:06 +00:00
|
|
|
}
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(db_->Flush(FlushOptions(), handles_[1]));
|
2014-10-21 18:23:06 +00:00
|
|
|
for (int i = 0; i < 1000; ++i) {
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), handles_[1], test::RandomKey(&rnd, 10),
|
|
|
|
test::RandomKey(&rnd, 10)));
|
2014-10-21 18:23:06 +00:00
|
|
|
}
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_OK(db_->Flush(FlushOptions(), handles_[1]));
|
2014-10-21 18:23:06 +00:00
|
|
|
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
2019-09-18 23:56:25 +00:00
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
2014-10-21 18:23:06 +00:00
|
|
|
ASSERT_EQ(2U, metadata.size());
|
|
|
|
ASSERT_EQ("new_cf", metadata[0].column_family_name);
|
|
|
|
ASSERT_EQ("new_cf", metadata[1].column_family_name);
|
|
|
|
auto old_file = metadata[0].smallest_seqno < metadata[1].smallest_seqno
|
|
|
|
? metadata[0].name
|
|
|
|
: metadata[1].name;
|
|
|
|
auto new_file = metadata[0].smallest_seqno > metadata[1].smallest_seqno
|
|
|
|
? metadata[0].name
|
|
|
|
: metadata[1].name;
|
2019-09-18 23:56:25 +00:00
|
|
|
ASSERT_TRUE(db_->DeleteFile(new_file).IsInvalidArgument());
|
|
|
|
ASSERT_OK(db_->DeleteFile(old_file));
|
2014-10-21 18:23:06 +00:00
|
|
|
|
|
|
|
{
|
2019-09-18 23:56:25 +00:00
|
|
|
std::unique_ptr<Iterator> itr(db_->NewIterator(ReadOptions(), handles_[1]));
|
2020-12-22 23:08:17 +00:00
|
|
|
ASSERT_OK(itr->status());
|
2014-10-21 18:23:06 +00:00
|
|
|
int count = 0;
|
|
|
|
for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {
|
|
|
|
ASSERT_OK(itr->status());
|
|
|
|
++count;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(count, 1000);
|
|
|
|
}
|
|
|
|
|
2019-09-18 23:56:25 +00:00
|
|
|
Close();
|
|
|
|
ReopenWithColumnFamilies({kDefaultColumnFamilyName, "new_cf"}, options);
|
2014-10-21 18:23:06 +00:00
|
|
|
|
|
|
|
{
|
2019-09-18 23:56:25 +00:00
|
|
|
std::unique_ptr<Iterator> itr(db_->NewIterator(ReadOptions(), handles_[1]));
|
2014-10-21 18:23:06 +00:00
|
|
|
int count = 0;
|
|
|
|
for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {
|
|
|
|
ASSERT_OK(itr->status());
|
|
|
|
++count;
|
|
|
|
}
|
|
|
|
ASSERT_EQ(count, 1000);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
2013-08-22 21:32:53 +00:00
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
2020-02-20 20:07:53 +00:00
|
|
|
ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
|
2015-03-17 21:08:00 +00:00
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
2019-09-18 23:56:25 +00:00
|
|
|
RegisterCustomObjects(argc, argv);
|
2015-03-17 21:08:00 +00:00
|
|
|
return RUN_ALL_TESTS();
|
2013-08-22 21:32:53 +00:00
|
|
|
}
|
|
|
|
|