mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-26 07:30:54 +00:00
69c471bd9b
Summary: Fixed two related race conditions in backup creation. (1) CreateNewBackup() uses DB::DisableFileDeletions() to prevent table files from being deleted while it is copying; however, the MANIFEST file could still rotate during this time. The fix is to stop deleting the old manifest in the rotation logic. It will be deleted safely later when PurgeObsoleteFiles() runs (can only happen when file deletions are enabled). (2) CreateNewBackup() did not account for the CURRENT file being mutable. This is significant because the files returned by GetLiveFiles() contain a particular manifest filename, but the manifest to which CURRENT refers can change at any time. This causes problems when CURRENT changes between the call to GetLiveFiles() and when it's copied to the backup directory. To workaround this, I manually forge a CURRENT file referring to the manifest filename returned in GetLiveFiles(). (2) also applies to the checkpointing code, so let me know if this approach is good and I'll make the same change there. Test Plan: new test for roll manifest during backup creation. running the test before this change: $ ./backupable_db_test --gtest_filter=BackupableDBTest.ChangeManifestDuringBackupCreation ... IO error: /tmp/rocksdbtest-9383/backupable_db/MANIFEST-000001: No such file or directory running the test after this change: $ ./backupable_db_test --gtest_filter=BackupableDBTest.ChangeManifestDuringBackupCreation ... [ RUN ] BackupableDBTest.ChangeManifestDuringBackupCreation [ OK ] BackupableDBTest.ChangeManifestDuringBackupCreation (2836 ms) Reviewers: IslamAbdelRahman, anthony, sdong Reviewed By: sdong Subscribers: dhruba, leveldb Differential Revision: https://reviews.facebook.net/D54711
388 lines
12 KiB
C++
388 lines
12 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under the BSD-style license found in the
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
#include "rocksdb/db.h"
|
|
#include "db/db_impl.h"
|
|
#include "db/filename.h"
|
|
#include "db/version_set.h"
|
|
#include "db/write_batch_internal.h"
|
|
#include "util/string_util.h"
|
|
#include "util/testharness.h"
|
|
#include "util/testutil.h"
|
|
#include "rocksdb/env.h"
|
|
#include "rocksdb/transaction_log.h"
|
|
#include <vector>
|
|
#include <stdlib.h>
|
|
#include <map>
|
|
#include <string>
|
|
|
|
namespace rocksdb {
|
|
|
|
class DeleteFileTest : public testing::Test {
|
|
public:
|
|
std::string dbname_;
|
|
Options options_;
|
|
DB* db_;
|
|
Env* env_;
|
|
int numlevels_;
|
|
|
|
DeleteFileTest() {
|
|
db_ = nullptr;
|
|
env_ = Env::Default();
|
|
options_.delete_obsolete_files_period_micros = 0; // always do full purge
|
|
options_.enable_thread_tracking = true;
|
|
options_.write_buffer_size = 1024*1024*1000;
|
|
options_.target_file_size_base = 1024*1024*1000;
|
|
options_.max_bytes_for_level_base = 1024*1024*1000;
|
|
options_.WAL_ttl_seconds = 300; // Used to test log files
|
|
options_.WAL_size_limit_MB = 1024; // Used to test log files
|
|
dbname_ = test::TmpDir() + "/deletefile_test";
|
|
options_.wal_dir = dbname_ + "/wal_files";
|
|
|
|
// clean up all the files that might have been there before
|
|
std::vector<std::string> old_files;
|
|
env_->GetChildren(dbname_, &old_files);
|
|
for (auto file : old_files) {
|
|
env_->DeleteFile(dbname_ + "/" + file);
|
|
}
|
|
env_->GetChildren(options_.wal_dir, &old_files);
|
|
for (auto file : old_files) {
|
|
env_->DeleteFile(options_.wal_dir + "/" + file);
|
|
}
|
|
|
|
DestroyDB(dbname_, options_);
|
|
numlevels_ = 7;
|
|
EXPECT_OK(ReopenDB(true));
|
|
}
|
|
|
|
Status ReopenDB(bool create) {
|
|
delete db_;
|
|
if (create) {
|
|
DestroyDB(dbname_, options_);
|
|
}
|
|
db_ = nullptr;
|
|
options_.create_if_missing = create;
|
|
return DB::Open(options_, dbname_, &db_);
|
|
}
|
|
|
|
void CloseDB() {
|
|
delete db_;
|
|
db_ = nullptr;
|
|
}
|
|
|
|
void AddKeys(int numkeys, int startkey = 0) {
|
|
WriteOptions options;
|
|
options.sync = false;
|
|
ReadOptions roptions;
|
|
for (int i = startkey; i < (numkeys + startkey) ; i++) {
|
|
std::string temp = ToString(i);
|
|
Slice key(temp);
|
|
Slice value(temp);
|
|
ASSERT_OK(db_->Put(options, key, value));
|
|
}
|
|
}
|
|
|
|
int numKeysInLevels(
|
|
std::vector<LiveFileMetaData> &metadata,
|
|
std::vector<int> *keysperlevel = nullptr) {
|
|
|
|
if (keysperlevel != nullptr) {
|
|
keysperlevel->resize(numlevels_);
|
|
}
|
|
|
|
int numKeys = 0;
|
|
for (size_t i = 0; i < metadata.size(); i++) {
|
|
int startkey = atoi(metadata[i].smallestkey.c_str());
|
|
int endkey = atoi(metadata[i].largestkey.c_str());
|
|
int numkeysinfile = (endkey - startkey + 1);
|
|
numKeys += numkeysinfile;
|
|
if (keysperlevel != nullptr) {
|
|
(*keysperlevel)[(int)metadata[i].level] += numkeysinfile;
|
|
}
|
|
fprintf(stderr, "level %d name %s smallest %s largest %s\n",
|
|
metadata[i].level, metadata[i].name.c_str(),
|
|
metadata[i].smallestkey.c_str(),
|
|
metadata[i].largestkey.c_str());
|
|
}
|
|
return numKeys;
|
|
}
|
|
|
|
void CreateTwoLevels() {
|
|
AddKeys(50000, 10000);
|
|
DBImpl* dbi = reinterpret_cast<DBImpl*>(db_);
|
|
ASSERT_OK(dbi->TEST_FlushMemTable());
|
|
ASSERT_OK(dbi->TEST_WaitForFlushMemTable());
|
|
for (int i = 0; i < 2; ++i) {
|
|
ASSERT_OK(dbi->TEST_CompactRange(i, nullptr, nullptr));
|
|
}
|
|
|
|
AddKeys(50000, 10000);
|
|
ASSERT_OK(dbi->TEST_FlushMemTable());
|
|
ASSERT_OK(dbi->TEST_WaitForFlushMemTable());
|
|
ASSERT_OK(dbi->TEST_CompactRange(0, nullptr, nullptr));
|
|
}
|
|
|
|
void CheckFileTypeCounts(std::string& dir,
|
|
int required_log,
|
|
int required_sst,
|
|
int required_manifest) {
|
|
std::vector<std::string> filenames;
|
|
env_->GetChildren(dir, &filenames);
|
|
|
|
int log_cnt = 0, sst_cnt = 0, manifest_cnt = 0;
|
|
for (auto file : filenames) {
|
|
uint64_t number;
|
|
FileType type;
|
|
if (ParseFileName(file, &number, &type)) {
|
|
log_cnt += (type == kLogFile);
|
|
sst_cnt += (type == kTableFile);
|
|
manifest_cnt += (type == kDescriptorFile);
|
|
}
|
|
}
|
|
ASSERT_EQ(required_log, log_cnt);
|
|
ASSERT_EQ(required_sst, sst_cnt);
|
|
ASSERT_EQ(required_manifest, manifest_cnt);
|
|
}
|
|
|
|
};
|
|
|
|
TEST_F(DeleteFileTest, AddKeysAndQueryLevels) {
|
|
CreateTwoLevels();
|
|
std::vector<LiveFileMetaData> metadata;
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
std::string level1file = "";
|
|
int level1keycount = 0;
|
|
std::string level2file = "";
|
|
int level2keycount = 0;
|
|
int level1index = 0;
|
|
int level2index = 1;
|
|
|
|
ASSERT_EQ((int)metadata.size(), 2);
|
|
if (metadata[0].level == 2) {
|
|
level1index = 1;
|
|
level2index = 0;
|
|
}
|
|
|
|
level1file = metadata[level1index].name;
|
|
int startkey = atoi(metadata[level1index].smallestkey.c_str());
|
|
int endkey = atoi(metadata[level1index].largestkey.c_str());
|
|
level1keycount = (endkey - startkey + 1);
|
|
level2file = metadata[level2index].name;
|
|
startkey = atoi(metadata[level2index].smallestkey.c_str());
|
|
endkey = atoi(metadata[level2index].largestkey.c_str());
|
|
level2keycount = (endkey - startkey + 1);
|
|
|
|
// COntrolled setup. Levels 1 and 2 should both have 50K files.
|
|
// This is a little fragile as it depends on the current
|
|
// compaction heuristics.
|
|
ASSERT_EQ(level1keycount, 50000);
|
|
ASSERT_EQ(level2keycount, 50000);
|
|
|
|
Status status = db_->DeleteFile("0.sst");
|
|
ASSERT_TRUE(status.IsInvalidArgument());
|
|
|
|
// intermediate level files cannot be deleted.
|
|
status = db_->DeleteFile(level1file);
|
|
ASSERT_TRUE(status.IsInvalidArgument());
|
|
|
|
// Lowest level file deletion should succeed.
|
|
ASSERT_OK(db_->DeleteFile(level2file));
|
|
|
|
CloseDB();
|
|
}
|
|
|
|
TEST_F(DeleteFileTest, PurgeObsoleteFilesTest) {
|
|
CreateTwoLevels();
|
|
// there should be only one (empty) log file because CreateTwoLevels()
|
|
// flushes the memtables to disk
|
|
CheckFileTypeCounts(options_.wal_dir, 1, 0, 0);
|
|
// 2 ssts, 1 manifest
|
|
CheckFileTypeCounts(dbname_, 0, 2, 1);
|
|
std::string first("0"), last("999999");
|
|
CompactRangeOptions compact_options;
|
|
compact_options.change_level = true;
|
|
compact_options.target_level = 2;
|
|
Slice first_slice(first), last_slice(last);
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
// 1 sst after compaction
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
// this time, we keep an iterator alive
|
|
ReopenDB(true);
|
|
Iterator *itr = 0;
|
|
CreateTwoLevels();
|
|
itr = db_->NewIterator(ReadOptions());
|
|
db_->CompactRange(compact_options, &first_slice, &last_slice);
|
|
// 3 sst after compaction with live iterator
|
|
CheckFileTypeCounts(dbname_, 0, 3, 1);
|
|
delete itr;
|
|
// 1 sst after iterator deletion
|
|
CheckFileTypeCounts(dbname_, 0, 1, 1);
|
|
|
|
CloseDB();
|
|
}
|
|
|
|
TEST_F(DeleteFileTest, DeleteFileWithIterator) {
|
|
CreateTwoLevels();
|
|
ReadOptions options;
|
|
Iterator* it = db_->NewIterator(options);
|
|
std::vector<LiveFileMetaData> metadata;
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
std::string level2file = "";
|
|
|
|
ASSERT_EQ((int)metadata.size(), 2);
|
|
if (metadata[0].level == 1) {
|
|
level2file = metadata[1].name;
|
|
} else {
|
|
level2file = metadata[0].name;
|
|
}
|
|
|
|
Status status = db_->DeleteFile(level2file);
|
|
fprintf(stdout, "Deletion status %s: %s\n",
|
|
level2file.c_str(), status.ToString().c_str());
|
|
ASSERT_TRUE(status.ok());
|
|
it->SeekToFirst();
|
|
int numKeysIterated = 0;
|
|
while(it->Valid()) {
|
|
numKeysIterated++;
|
|
it->Next();
|
|
}
|
|
ASSERT_EQ(numKeysIterated, 50000);
|
|
delete it;
|
|
CloseDB();
|
|
}
|
|
|
|
TEST_F(DeleteFileTest, DeleteLogFiles) {
|
|
AddKeys(10, 0);
|
|
VectorLogPtr logfiles;
|
|
db_->GetSortedWalFiles(logfiles);
|
|
ASSERT_GT(logfiles.size(), 0UL);
|
|
// Take the last log file which is expected to be alive and try to delete it
|
|
// Should not succeed because live logs are not allowed to be deleted
|
|
std::unique_ptr<LogFile> alive_log = std::move(logfiles.back());
|
|
ASSERT_EQ(alive_log->Type(), kAliveLogFile);
|
|
ASSERT_OK(env_->FileExists(options_.wal_dir + "/" + alive_log->PathName()));
|
|
fprintf(stdout, "Deleting alive log file %s\n",
|
|
alive_log->PathName().c_str());
|
|
ASSERT_TRUE(!db_->DeleteFile(alive_log->PathName()).ok());
|
|
ASSERT_OK(env_->FileExists(options_.wal_dir + "/" + alive_log->PathName()));
|
|
logfiles.clear();
|
|
|
|
// Call Flush to bring about a new working log file and add more keys
|
|
// Call Flush again to flush out memtable and move alive log to archived log
|
|
// and try to delete the archived log file
|
|
FlushOptions fopts;
|
|
db_->Flush(fopts);
|
|
AddKeys(10, 0);
|
|
db_->Flush(fopts);
|
|
db_->GetSortedWalFiles(logfiles);
|
|
ASSERT_GT(logfiles.size(), 0UL);
|
|
std::unique_ptr<LogFile> archived_log = std::move(logfiles.front());
|
|
ASSERT_EQ(archived_log->Type(), kArchivedLogFile);
|
|
ASSERT_OK(
|
|
env_->FileExists(options_.wal_dir + "/" + archived_log->PathName()));
|
|
fprintf(stdout, "Deleting archived log file %s\n",
|
|
archived_log->PathName().c_str());
|
|
ASSERT_OK(db_->DeleteFile(archived_log->PathName()));
|
|
ASSERT_EQ(Status::NotFound(), env_->FileExists(options_.wal_dir + "/" +
|
|
archived_log->PathName()));
|
|
CloseDB();
|
|
}
|
|
|
|
TEST_F(DeleteFileTest, DeleteNonDefaultColumnFamily) {
|
|
CloseDB();
|
|
DBOptions db_options;
|
|
db_options.create_if_missing = true;
|
|
db_options.create_missing_column_families = true;
|
|
std::vector<ColumnFamilyDescriptor> column_families;
|
|
column_families.emplace_back();
|
|
column_families.emplace_back("new_cf", ColumnFamilyOptions());
|
|
|
|
std::vector<rocksdb::ColumnFamilyHandle*> handles;
|
|
rocksdb::DB* db;
|
|
ASSERT_OK(DB::Open(db_options, dbname_, column_families, &handles, &db));
|
|
|
|
Random rnd(5);
|
|
for (int i = 0; i < 1000; ++i) {
|
|
ASSERT_OK(db->Put(WriteOptions(), handles[1], test::RandomKey(&rnd, 10),
|
|
test::RandomKey(&rnd, 10)));
|
|
}
|
|
ASSERT_OK(db->Flush(FlushOptions(), handles[1]));
|
|
for (int i = 0; i < 1000; ++i) {
|
|
ASSERT_OK(db->Put(WriteOptions(), handles[1], test::RandomKey(&rnd, 10),
|
|
test::RandomKey(&rnd, 10)));
|
|
}
|
|
ASSERT_OK(db->Flush(FlushOptions(), handles[1]));
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
db->GetLiveFilesMetaData(&metadata);
|
|
ASSERT_EQ(2U, metadata.size());
|
|
ASSERT_EQ("new_cf", metadata[0].column_family_name);
|
|
ASSERT_EQ("new_cf", metadata[1].column_family_name);
|
|
auto old_file = metadata[0].smallest_seqno < metadata[1].smallest_seqno
|
|
? metadata[0].name
|
|
: metadata[1].name;
|
|
auto new_file = metadata[0].smallest_seqno > metadata[1].smallest_seqno
|
|
? metadata[0].name
|
|
: metadata[1].name;
|
|
ASSERT_TRUE(db->DeleteFile(new_file).IsInvalidArgument());
|
|
ASSERT_OK(db->DeleteFile(old_file));
|
|
|
|
{
|
|
std::unique_ptr<Iterator> itr(db->NewIterator(ReadOptions(), handles[1]));
|
|
int count = 0;
|
|
for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {
|
|
ASSERT_OK(itr->status());
|
|
++count;
|
|
}
|
|
ASSERT_EQ(count, 1000);
|
|
}
|
|
|
|
delete handles[0];
|
|
delete handles[1];
|
|
delete db;
|
|
|
|
ASSERT_OK(DB::Open(db_options, dbname_, column_families, &handles, &db));
|
|
{
|
|
std::unique_ptr<Iterator> itr(db->NewIterator(ReadOptions(), handles[1]));
|
|
int count = 0;
|
|
for (itr->SeekToFirst(); itr->Valid(); itr->Next()) {
|
|
ASSERT_OK(itr->status());
|
|
++count;
|
|
}
|
|
ASSERT_EQ(count, 1000);
|
|
}
|
|
|
|
delete handles[0];
|
|
delete handles[1];
|
|
delete db;
|
|
}
|
|
|
|
} //namespace rocksdb
|
|
|
|
int main(int argc, char** argv) {
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
return RUN_ALL_TESTS();
|
|
}
|
|
|
|
#else
|
|
#include <stdio.h>
|
|
|
|
int main(int argc, char** argv) {
|
|
fprintf(stderr,
|
|
"SKIPPED as DBImpl::DeleteFile is not supported in ROCKSDB_LITE\n");
|
|
return 0;
|
|
}
|
|
|
|
#endif // !ROCKSDB_LITE
|