2016-02-09 23:12:00 +00:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-15 23:03:42 +00:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2013-10-16 21:59:46 +00:00
|
|
|
//
|
2011-03-18 22:37:00 +00:00
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
2015-09-14 23:44:35 +00:00
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
|
2011-03-18 22:37:00 +00:00
|
|
|
#include <errno.h>
|
|
|
|
#include <fcntl.h>
|
|
|
|
#include <sys/stat.h>
|
|
|
|
#include <sys/types.h>
|
2020-07-09 21:33:42 +00:00
|
|
|
|
2019-08-16 23:40:09 +00:00
|
|
|
#include <cinttypes>
|
2020-07-09 21:33:42 +00:00
|
|
|
|
2019-05-31 18:52:59 +00:00
|
|
|
#include "db/db_impl/db_impl.h"
|
2019-08-16 23:40:09 +00:00
|
|
|
#include "db/db_test_util.h"
|
2017-04-04 01:27:24 +00:00
|
|
|
#include "db/log_format.h"
|
|
|
|
#include "db/version_set.h"
|
Introduce a new storage specific Env API (#5761)
Summary:
The current Env API encompasses both storage/file operations, as well as OS related operations. Most of the APIs return a Status, which does not have enough metadata about an error, such as whether its retry-able or not, scope (i.e fault domain) of the error etc., that may be required in order to properly handle a storage error. The file APIs also do not provide enough control over the IO SLA, such as timeout, prioritization, hinting about placement and redundancy etc.
This PR separates out the file/storage APIs from Env into a new FileSystem class. The APIs are updated to return an IOStatus with metadata about the error, as well as to take an IOOptions structure as input in order to allow more control over the IO.
The user can set both ```options.env``` and ```options.file_system``` to specify that RocksDB should use the former for OS related operations and the latter for storage operations. Internally, a ```CompositeEnvWrapper``` has been introduced that inherits from ```Env``` and redirects individual methods to either an ```Env``` implementation or the ```FileSystem``` as appropriate. When options are sanitized during ```DB::Open```, ```options.env``` is replaced with a newly allocated ```CompositeEnvWrapper``` instance if both env and file_system have been specified. This way, the rest of the RocksDB code can continue to function as before.
This PR also ports PosixEnv to the new API by splitting it into two - PosixEnv and PosixFileSystem. PosixEnv is defined as a sub-class of CompositeEnvWrapper, and threading/time functions are overridden with Posix specific implementations in order to avoid an extra level of indirection.
The ```CompositeEnvWrapper``` translates ```IOStatus``` return code to ```Status```, and sets the severity to ```kSoftError``` if the io_status is retryable. The error handling code in RocksDB can then recover the DB automatically.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5761
Differential Revision: D18868376
Pulled By: anand1976
fbshipit-source-id: 39efe18a162ea746fabac6360ff529baba48486f
2019-12-13 22:47:08 +00:00
|
|
|
#include "env/composite_env_wrapper.h"
|
2019-05-30 03:44:08 +00:00
|
|
|
#include "file/filename.h"
|
2013-08-23 15:38:13 +00:00
|
|
|
#include "rocksdb/cache.h"
|
2017-08-09 22:49:40 +00:00
|
|
|
#include "rocksdb/convenience.h"
|
2020-07-09 21:33:42 +00:00
|
|
|
#include "rocksdb/db.h"
|
2013-08-23 15:38:13 +00:00
|
|
|
#include "rocksdb/env.h"
|
2013-10-29 00:54:09 +00:00
|
|
|
#include "rocksdb/table.h"
|
2013-08-23 15:38:13 +00:00
|
|
|
#include "rocksdb/write_batch.h"
|
2019-05-30 21:47:29 +00:00
|
|
|
#include "table/block_based/block_based_table_builder.h"
|
2018-12-21 05:57:18 +00:00
|
|
|
#include "table/meta_blocks.h"
|
2020-07-22 18:03:29 +00:00
|
|
|
#include "table/mock_table.h"
|
2019-05-30 18:21:38 +00:00
|
|
|
#include "test_util/testharness.h"
|
|
|
|
#include "test_util/testutil.h"
|
2020-07-09 21:33:42 +00:00
|
|
|
#include "util/random.h"
|
2019-05-31 00:39:43 +00:00
|
|
|
#include "util/string_util.h"
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
static const int kValueSize = 1000;
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
class CorruptionTest : public testing::Test {
|
2011-03-18 22:37:00 +00:00
|
|
|
public:
|
|
|
|
test::ErrorEnv env_;
|
|
|
|
std::string dbname_;
|
2018-11-09 19:17:34 +00:00
|
|
|
std::shared_ptr<Cache> tiny_cache_;
|
2011-03-18 22:37:00 +00:00
|
|
|
Options options_;
|
|
|
|
DB* db_;
|
|
|
|
|
2011-06-22 02:36:45 +00:00
|
|
|
CorruptionTest() {
|
2017-01-27 14:35:41 +00:00
|
|
|
// If LRU cache shard bit is smaller than 2 (or -1 which will automatically
|
|
|
|
// set it to 0), test SequenceNumberRecovery will fail, likely because of a
|
|
|
|
// bug in recovery code. Keep it 4 for now to make the test passes.
|
2017-01-27 19:15:02 +00:00
|
|
|
tiny_cache_ = NewLRUCache(100, 4);
|
2016-04-08 00:40:42 +00:00
|
|
|
options_.wal_recovery_mode = WALRecoveryMode::kTolerateCorruptedTailRecords;
|
2011-03-18 22:37:00 +00:00
|
|
|
options_.env = &env_;
|
2018-07-14 00:18:39 +00:00
|
|
|
dbname_ = test::PerThreadDBPath("corruption_test");
|
2011-03-18 22:37:00 +00:00
|
|
|
DestroyDB(dbname_, options_);
|
|
|
|
|
2013-03-01 02:04:58 +00:00
|
|
|
db_ = nullptr;
|
2011-03-18 22:37:00 +00:00
|
|
|
options_.create_if_missing = true;
|
2014-08-25 21:22:05 +00:00
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.block_size_deviation = 0; // make unit test pass for now
|
|
|
|
options_.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
2011-03-18 22:37:00 +00:00
|
|
|
Reopen();
|
|
|
|
options_.create_if_missing = false;
|
|
|
|
}
|
|
|
|
|
2019-02-14 21:52:47 +00:00
|
|
|
~CorruptionTest() override {
|
|
|
|
delete db_;
|
2020-08-13 02:28:28 +00:00
|
|
|
db_ = nullptr;
|
2019-02-14 21:52:47 +00:00
|
|
|
DestroyDB(dbname_, Options());
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2015-07-01 23:13:49 +00:00
|
|
|
void CloseDb() {
|
|
|
|
delete db_;
|
|
|
|
db_ = nullptr;
|
|
|
|
}
|
|
|
|
|
2013-03-01 02:04:58 +00:00
|
|
|
Status TryReopen(Options* options = nullptr) {
|
2011-03-18 22:37:00 +00:00
|
|
|
delete db_;
|
2013-03-01 02:04:58 +00:00
|
|
|
db_ = nullptr;
|
2011-03-18 22:37:00 +00:00
|
|
|
Options opt = (options ? *options : options_);
|
2019-08-16 23:40:09 +00:00
|
|
|
if (opt.env == Options().env) {
|
|
|
|
// If env is not overridden, replace it with ErrorEnv.
|
|
|
|
// Otherwise, the test already uses a non-default Env.
|
|
|
|
opt.env = &env_;
|
|
|
|
}
|
Make arena block size configurable
Summary:
Add an option for arena block size, default value 4096 bytes. Arena will allocate blocks with such size.
I am not sure about passing parameter to skiplist in the new virtualized framework, though I talked to Jim a bit. So add Jim as reviewer.
Test Plan:
new unit test, I am running db_test.
For passing paramter from configured option to Arena, I tried tests like:
TEST(DBTest, Arena_Option) {
std::string dbname = test::TmpDir() + "/db_arena_option_test";
DestroyDB(dbname, Options());
DB* db = nullptr;
Options opts;
opts.create_if_missing = true;
opts.arena_block_size = 1000000; // tested 99, 999999
Status s = DB::Open(opts, dbname, &db);
db->Put(WriteOptions(), "a", "123");
}
and printed some debug info. The results look good. Any suggestion for such a unit-test?
Reviewers: haobo, dhruba, emayanke, jpaton
Reviewed By: dhruba
CC: leveldb, zshao
Differential Revision: https://reviews.facebook.net/D11799
2013-07-31 19:42:23 +00:00
|
|
|
opt.arena_block_size = 4096;
|
2014-08-25 21:22:05 +00:00
|
|
|
BlockBasedTableOptions table_options;
|
|
|
|
table_options.block_cache = tiny_cache_;
|
|
|
|
table_options.block_size_deviation = 0;
|
|
|
|
opt.table_factory.reset(NewBlockBasedTableFactory(table_options));
|
2011-03-18 22:37:00 +00:00
|
|
|
return DB::Open(opt, dbname_, &db_);
|
|
|
|
}
|
|
|
|
|
2013-03-01 02:04:58 +00:00
|
|
|
void Reopen(Options* options = nullptr) {
|
2011-03-18 22:37:00 +00:00
|
|
|
ASSERT_OK(TryReopen(options));
|
|
|
|
}
|
|
|
|
|
|
|
|
void RepairDB() {
|
|
|
|
delete db_;
|
2013-03-01 02:04:58 +00:00
|
|
|
db_ = nullptr;
|
2020-02-20 20:07:53 +00:00
|
|
|
ASSERT_OK(::ROCKSDB_NAMESPACE::RepairDB(dbname_, options_));
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2015-10-05 21:46:28 +00:00
|
|
|
void Build(int n, int flush_every = 0) {
|
2011-03-18 22:37:00 +00:00
|
|
|
std::string key_space, value_space;
|
|
|
|
WriteBatch batch;
|
|
|
|
for (int i = 0; i < n; i++) {
|
2015-10-05 21:46:28 +00:00
|
|
|
if (flush_every != 0 && i != 0 && i % flush_every == 0) {
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2015-10-05 21:46:28 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
|
|
|
}
|
2011-03-18 22:37:00 +00:00
|
|
|
//if ((i % 100) == 0) fprintf(stderr, "@ %d of %d\n", i, n);
|
|
|
|
Slice key = Key(i, &key_space);
|
|
|
|
batch.Clear();
|
|
|
|
batch.Put(key, Value(i, &value_space));
|
|
|
|
ASSERT_OK(db_->Write(WriteOptions(), &batch));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void Check(int min_expected, int max_expected) {
|
2015-11-19 19:47:12 +00:00
|
|
|
uint64_t next_expected = 0;
|
|
|
|
uint64_t missed = 0;
|
2011-03-18 22:37:00 +00:00
|
|
|
int bad_keys = 0;
|
|
|
|
int bad_values = 0;
|
|
|
|
int correct = 0;
|
|
|
|
std::string value_space;
|
2014-02-01 01:16:30 +00:00
|
|
|
// Do not verify checksums. If we verify checksums then the
|
|
|
|
// db itself will raise errors because data is corrupted.
|
|
|
|
// Instead, we want the reads to be successful and this test
|
|
|
|
// will detect whether the appropriate corruptions have
|
2015-04-25 09:14:27 +00:00
|
|
|
// occurred.
|
2014-02-01 01:16:30 +00:00
|
|
|
Iterator* iter = db_->NewIterator(ReadOptions(false, true));
|
2011-03-18 22:37:00 +00:00
|
|
|
for (iter->SeekToFirst(); iter->Valid(); iter->Next()) {
|
|
|
|
uint64_t key;
|
|
|
|
Slice in(iter->key());
|
|
|
|
if (!ConsumeDecimalNumber(&in, &key) ||
|
|
|
|
!in.empty() ||
|
|
|
|
key < next_expected) {
|
|
|
|
bad_keys++;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
missed += (key - next_expected);
|
2015-11-19 19:47:12 +00:00
|
|
|
next_expected = key + 1;
|
2014-11-11 21:47:22 +00:00
|
|
|
if (iter->value() != Value(static_cast<int>(key), &value_space)) {
|
2011-03-18 22:37:00 +00:00
|
|
|
bad_values++;
|
|
|
|
} else {
|
|
|
|
correct++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
delete iter;
|
|
|
|
|
|
|
|
fprintf(stderr,
|
2015-11-19 19:47:12 +00:00
|
|
|
"expected=%d..%d; got=%d; bad_keys=%d; bad_values=%d; missed=%llu\n",
|
|
|
|
min_expected, max_expected, correct, bad_keys, bad_values,
|
|
|
|
static_cast<unsigned long long>(missed));
|
2011-03-18 22:37:00 +00:00
|
|
|
ASSERT_LE(min_expected, correct);
|
|
|
|
ASSERT_GE(max_expected, correct);
|
|
|
|
}
|
|
|
|
|
2014-04-24 18:13:28 +00:00
|
|
|
void Corrupt(FileType filetype, int offset, int bytes_to_corrupt) {
|
|
|
|
// Pick file to corrupt
|
|
|
|
std::vector<std::string> filenames;
|
|
|
|
ASSERT_OK(env_.GetChildren(dbname_, &filenames));
|
|
|
|
uint64_t number;
|
|
|
|
FileType type;
|
|
|
|
std::string fname;
|
|
|
|
int picked_number = -1;
|
2015-12-15 23:26:20 +00:00
|
|
|
for (size_t i = 0; i < filenames.size(); i++) {
|
2014-04-24 18:13:28 +00:00
|
|
|
if (ParseFileName(filenames[i], &number, &type) &&
|
|
|
|
type == filetype &&
|
|
|
|
static_cast<int>(number) > picked_number) { // Pick latest file
|
|
|
|
fname = dbname_ + "/" + filenames[i];
|
2014-11-11 21:47:22 +00:00
|
|
|
picked_number = static_cast<int>(number);
|
2014-04-24 18:13:28 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
ASSERT_TRUE(!fname.empty()) << filetype;
|
|
|
|
|
2020-06-05 18:06:26 +00:00
|
|
|
test::CorruptFile(fname, offset, bytes_to_corrupt);
|
2014-04-24 18:13:28 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// corrupts exactly one file at level `level`. if no file found at level,
|
|
|
|
// asserts
|
|
|
|
void CorruptTableFileAtLevel(int level, int offset, int bytes_to_corrupt) {
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
for (const auto& m : metadata) {
|
|
|
|
if (m.level == level) {
|
2020-06-05 18:06:26 +00:00
|
|
|
test::CorruptFile(dbname_ + "/" + m.name, offset, bytes_to_corrupt);
|
2014-04-24 18:13:28 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
2017-07-17 04:23:33 +00:00
|
|
|
FAIL() << "no file found at level";
|
2014-04-24 18:13:28 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2011-04-12 19:38:58 +00:00
|
|
|
int Property(const std::string& name) {
|
|
|
|
std::string property;
|
|
|
|
int result;
|
|
|
|
if (db_->GetProperty(name, &property) &&
|
|
|
|
sscanf(property.c_str(), "%d", &result) == 1) {
|
|
|
|
return result;
|
|
|
|
} else {
|
|
|
|
return -1;
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return the ith key
|
|
|
|
Slice Key(int i, std::string* storage) {
|
|
|
|
char buf[100];
|
|
|
|
snprintf(buf, sizeof(buf), "%016d", i);
|
|
|
|
storage->assign(buf, strlen(buf));
|
|
|
|
return Slice(*storage);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return the value to associate with the specified key
|
|
|
|
Slice Value(int k, std::string* storage) {
|
2015-11-06 16:07:08 +00:00
|
|
|
if (k == 0) {
|
|
|
|
// Ugh. Random seed of 0 used to produce no entropy. This code
|
|
|
|
// preserves the implementation that was in place when all of the
|
|
|
|
// magic values in this file were picked.
|
|
|
|
*storage = std::string(kValueSize, ' ');
|
|
|
|
} else {
|
|
|
|
Random r(k);
|
2020-07-09 21:33:42 +00:00
|
|
|
*storage = r.RandomString(kValueSize);
|
2015-11-06 16:07:08 +00:00
|
|
|
}
|
2020-07-09 21:33:42 +00:00
|
|
|
return Slice(*storage);
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, Recovery) {
|
2011-03-21 19:40:57 +00:00
|
|
|
Build(100);
|
|
|
|
Check(100, 100);
|
2015-07-01 23:13:49 +00:00
|
|
|
#ifdef OS_WIN
|
|
|
|
// On Wndows OS Disk cache does not behave properly
|
|
|
|
// We do not call FlushBuffers on every Flush. If we do not close
|
|
|
|
// the log file prior to the corruption we end up with the first
|
|
|
|
// block not corrupted but only the second. However, under the debugger
|
|
|
|
// things work just fine but never pass when running normally
|
|
|
|
// For that reason people may want to run with unbuffered I/O. That option
|
|
|
|
// is not available for WAL though.
|
|
|
|
CloseDb();
|
|
|
|
#endif
|
2011-03-18 22:37:00 +00:00
|
|
|
Corrupt(kLogFile, 19, 1); // WriteBatch tag for first record
|
2011-03-21 19:40:57 +00:00
|
|
|
Corrupt(kLogFile, log::kBlockSize + 1000, 1); // Somewhere in second block
|
2015-01-05 18:26:34 +00:00
|
|
|
ASSERT_TRUE(!TryReopen().ok());
|
|
|
|
options_.paranoid_checks = false;
|
|
|
|
Reopen(&options_);
|
2011-03-21 19:40:57 +00:00
|
|
|
|
|
|
|
// The 64 records in the first two log blocks are completely lost.
|
|
|
|
Check(36, 36);
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, RecoverWriteError) {
|
2011-03-18 22:37:00 +00:00
|
|
|
env_.writable_file_error_ = true;
|
|
|
|
Status s = TryReopen();
|
|
|
|
ASSERT_TRUE(!s.ok());
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, NewFileErrorDuringWrite) {
|
2011-03-18 22:37:00 +00:00
|
|
|
// Do enough writing to force minor compaction
|
|
|
|
env_.writable_file_error_ = true;
|
2014-11-11 21:47:22 +00:00
|
|
|
const int num =
|
|
|
|
static_cast<int>(3 + (Options().write_buffer_size / kValueSize));
|
2011-03-18 22:37:00 +00:00
|
|
|
std::string value_storage;
|
|
|
|
Status s;
|
2013-10-28 19:36:02 +00:00
|
|
|
bool failed = false;
|
|
|
|
for (int i = 0; i < num; i++) {
|
2011-03-18 22:37:00 +00:00
|
|
|
WriteBatch batch;
|
|
|
|
batch.Put("a", Value(100, &value_storage));
|
|
|
|
s = db_->Write(WriteOptions(), &batch);
|
2013-10-28 19:36:02 +00:00
|
|
|
if (!s.ok()) {
|
|
|
|
failed = true;
|
|
|
|
}
|
|
|
|
ASSERT_TRUE(!failed || !s.ok());
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
ASSERT_TRUE(!s.ok());
|
|
|
|
ASSERT_GE(env_.num_writable_file_errors_, 1);
|
|
|
|
env_.writable_file_error_ = false;
|
|
|
|
Reopen();
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, TableFile) {
|
2011-03-18 22:37:00 +00:00
|
|
|
Build(100);
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2013-03-01 02:04:58 +00:00
|
|
|
dbi->TEST_CompactRange(0, nullptr, nullptr);
|
|
|
|
dbi->TEST_CompactRange(1, nullptr, nullptr);
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
Corrupt(kTableFile, 100, 1);
|
|
|
|
Check(99, 99);
|
2017-08-09 22:49:40 +00:00
|
|
|
ASSERT_NOK(dbi->VerifyChecksum());
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2019-08-16 23:40:09 +00:00
|
|
|
TEST_F(CorruptionTest, VerifyChecksumReadahead) {
|
|
|
|
Options options;
|
|
|
|
SpecialEnv senv(Env::Default());
|
|
|
|
options.env = &senv;
|
|
|
|
// Disable block cache as we are going to check checksum for
|
|
|
|
// the same file twice and measure number of reads.
|
|
|
|
BlockBasedTableOptions table_options_no_bc;
|
|
|
|
table_options_no_bc.no_block_cache = true;
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory(table_options_no_bc));
|
|
|
|
|
|
|
|
Reopen(&options);
|
|
|
|
|
|
|
|
Build(10000);
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2019-08-16 23:40:09 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
|
|
|
dbi->TEST_CompactRange(0, nullptr, nullptr);
|
|
|
|
dbi->TEST_CompactRange(1, nullptr, nullptr);
|
|
|
|
|
|
|
|
senv.count_random_reads_ = true;
|
|
|
|
senv.random_read_counter_.Reset();
|
|
|
|
ASSERT_OK(dbi->VerifyChecksum());
|
|
|
|
|
|
|
|
// Make sure the counter is enabled.
|
|
|
|
ASSERT_GT(senv.random_read_counter_.Read(), 0);
|
|
|
|
|
|
|
|
// The SST file is about 10MB. Default readahead size is 256KB.
|
|
|
|
// Give a conservative 20 reads for metadata blocks, The number
|
|
|
|
// of random reads should be within 10 MB / 256KB + 20 = 60.
|
|
|
|
ASSERT_LT(senv.random_read_counter_.Read(), 60);
|
|
|
|
|
|
|
|
senv.random_read_bytes_counter_ = 0;
|
|
|
|
ReadOptions ro;
|
|
|
|
ro.readahead_size = size_t{32 * 1024};
|
|
|
|
ASSERT_OK(dbi->VerifyChecksum(ro));
|
|
|
|
// The SST file is about 10MB. We set readahead size to 32KB.
|
|
|
|
// Give 0 to 20 reads for metadata blocks, and allow real read
|
|
|
|
// to range from 24KB to 48KB. The lower bound would be:
|
|
|
|
// 10MB / 48KB + 0 = 213
|
|
|
|
// The higher bound is
|
|
|
|
// 10MB / 24KB + 20 = 447.
|
|
|
|
ASSERT_GE(senv.random_read_counter_.Read(), 213);
|
|
|
|
ASSERT_LE(senv.random_read_counter_.Read(), 447);
|
|
|
|
|
2019-10-21 18:37:09 +00:00
|
|
|
// Test readahead shouldn't break mmap mode (where it should be
|
|
|
|
// disabled).
|
|
|
|
options.allow_mmap_reads = true;
|
|
|
|
Reopen(&options);
|
|
|
|
dbi = static_cast<DBImpl*>(db_);
|
|
|
|
ASSERT_OK(dbi->VerifyChecksum(ro));
|
|
|
|
|
2019-08-16 23:40:09 +00:00
|
|
|
CloseDb();
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, TableFileIndexData) {
|
2015-10-05 21:46:28 +00:00
|
|
|
Options options;
|
|
|
|
// very big, we'll trigger flushes manually
|
|
|
|
options.write_buffer_size = 100 * 1024 * 1024;
|
|
|
|
Reopen(&options);
|
|
|
|
// build 2 tables, flush at 5000
|
|
|
|
Build(10000, 5000);
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2015-10-05 21:46:28 +00:00
|
|
|
// corrupt an index block of an entire file
|
2011-04-12 19:38:58 +00:00
|
|
|
Corrupt(kTableFile, -2000, 500);
|
2020-02-14 16:15:24 +00:00
|
|
|
options.paranoid_checks = false;
|
|
|
|
Reopen(&options);
|
2020-07-03 02:24:25 +00:00
|
|
|
dbi = static_cast_with_check<DBImpl>(db_);
|
Change and clarify the relationship between Valid(), status() and Seek*() for all iterators. Also fix some bugs
Summary:
Before this PR, Iterator/InternalIterator may simultaneously have non-ok status() and Valid() = true. That state means that the last operation failed, but the iterator is nevertheless positioned on some unspecified record. Likely intended uses of that are:
* If some sst files are corrupted, a normal iterator can be used to read the data from files that are not corrupted.
* When using read_tier = kBlockCacheTier, read the data that's in block cache, skipping over the data that is not.
However, this behavior wasn't documented well (and until recently the wiki on github had misleading incorrect information). In the code there's a lot of confusion about the relationship between status() and Valid(), and about whether Seek()/SeekToLast()/etc reset the status or not. There were a number of bugs caused by this confusion, both inside rocksdb and in the code that uses rocksdb (including ours).
This PR changes the convention to:
* If status() is not ok, Valid() always returns false.
* Any seek operation resets status. (Before the PR, it depended on iterator type and on particular error.)
This does sacrifice the two use cases listed above, but siying said it's ok.
Overview of the changes:
* A commit that adds missing status checks in MergingIterator. This fixes a bug that actually affects us, and we need it fixed. `DBIteratorTest.NonBlockingIterationBugRepro` explains the scenario.
* Changes to lots of iterator types to make all of them conform to the new convention. Some bug fixes along the way. By far the biggest changes are in DBIter, which is a big messy piece of code; I tried to make it less big and messy but mostly failed.
* A stress-test for DBIter, to gain some confidence that I didn't break it. It does a few million random operations on the iterator, while occasionally modifying the underlying data (like ForwardIterator does) and occasionally returning non-ok status from internal iterator.
To find the iterator types that needed changes I searched for "public .*Iterator" in the code. Here's an overview of all 27 iterator types:
Iterators that didn't need changes:
* status() is always ok(), or Valid() is always false: MemTableIterator, ModelIter, TestIterator, KVIter (2 classes with this name anonymous namespaces), LoggingForwardVectorIterator, VectorIterator, MockTableIterator, EmptyIterator, EmptyInternalIterator.
* Thin wrappers that always pass through Valid() and status(): ArenaWrappedDBIter, TtlIterator, InternalIteratorFromIterator.
Iterators with changes (see inline comments for details):
* DBIter - an overhaul:
- It used to silently skip corrupted keys (`FindParseableKey()`), which seems dangerous. This PR makes it just stop immediately after encountering a corrupted key, just like it would for other kinds of corruption. Let me know if there was actually some deeper meaning in this behavior and I should put it back.
- It had a few code paths silently discarding subiterator's status. The stress test caught a few.
- The backwards iteration code path was expecting the internal iterator's set of keys to be immutable. It's probably always true in practice at the moment, since ForwardIterator doesn't support backwards iteration, but this PR fixes it anyway. See added DBIteratorTest.ReverseToForwardBug for an example.
- Some parts of backwards iteration code path even did things like `assert(iter_->Valid())` after a seek, which is never a safe assumption.
- It used to not reset status on seek for some types of errors.
- Some simplifications and better comments.
- Some things got more complicated from the added error handling. I'm open to ideas for how to make it nicer.
* MergingIterator - check status after every operation on every subiterator, and in some places assert that valid subiterators have ok status.
* ForwardIterator - changed to the new convention, also slightly simplified.
* ForwardLevelIterator - fixed some bugs and simplified.
* LevelIterator - simplified.
* TwoLevelIterator - changed to the new convention. Also fixed a bug that would make SeekForPrev() sometimes silently ignore errors from first_level_iter_.
* BlockBasedTableIterator - minor changes.
* BlockIter - replaced `SetStatus()` with `Invalidate()` to make sure non-ok BlockIter is always invalid.
* PlainTableIterator - some seeks used to not reset status.
* CuckooTableIterator - tiny code cleanup.
* ManagedIterator - fixed some bugs.
* BaseDeltaIterator - changed to the new convention and fixed a bug.
* BlobDBIterator - seeks used to not reset status.
* KeyConvertingIterator - some small change.
Closes https://github.com/facebook/rocksdb/pull/3810
Differential Revision: D7888019
Pulled By: al13n321
fbshipit-source-id: 4aaf6d3421c545d16722a815b2fa2e7912bc851d
2018-05-17 09:44:14 +00:00
|
|
|
// one full file may be readable, since only one was corrupted
|
2015-10-05 21:46:28 +00:00
|
|
|
// the other file should be fully non-readable, since index was corrupted
|
Change and clarify the relationship between Valid(), status() and Seek*() for all iterators. Also fix some bugs
Summary:
Before this PR, Iterator/InternalIterator may simultaneously have non-ok status() and Valid() = true. That state means that the last operation failed, but the iterator is nevertheless positioned on some unspecified record. Likely intended uses of that are:
* If some sst files are corrupted, a normal iterator can be used to read the data from files that are not corrupted.
* When using read_tier = kBlockCacheTier, read the data that's in block cache, skipping over the data that is not.
However, this behavior wasn't documented well (and until recently the wiki on github had misleading incorrect information). In the code there's a lot of confusion about the relationship between status() and Valid(), and about whether Seek()/SeekToLast()/etc reset the status or not. There were a number of bugs caused by this confusion, both inside rocksdb and in the code that uses rocksdb (including ours).
This PR changes the convention to:
* If status() is not ok, Valid() always returns false.
* Any seek operation resets status. (Before the PR, it depended on iterator type and on particular error.)
This does sacrifice the two use cases listed above, but siying said it's ok.
Overview of the changes:
* A commit that adds missing status checks in MergingIterator. This fixes a bug that actually affects us, and we need it fixed. `DBIteratorTest.NonBlockingIterationBugRepro` explains the scenario.
* Changes to lots of iterator types to make all of them conform to the new convention. Some bug fixes along the way. By far the biggest changes are in DBIter, which is a big messy piece of code; I tried to make it less big and messy but mostly failed.
* A stress-test for DBIter, to gain some confidence that I didn't break it. It does a few million random operations on the iterator, while occasionally modifying the underlying data (like ForwardIterator does) and occasionally returning non-ok status from internal iterator.
To find the iterator types that needed changes I searched for "public .*Iterator" in the code. Here's an overview of all 27 iterator types:
Iterators that didn't need changes:
* status() is always ok(), or Valid() is always false: MemTableIterator, ModelIter, TestIterator, KVIter (2 classes with this name anonymous namespaces), LoggingForwardVectorIterator, VectorIterator, MockTableIterator, EmptyIterator, EmptyInternalIterator.
* Thin wrappers that always pass through Valid() and status(): ArenaWrappedDBIter, TtlIterator, InternalIteratorFromIterator.
Iterators with changes (see inline comments for details):
* DBIter - an overhaul:
- It used to silently skip corrupted keys (`FindParseableKey()`), which seems dangerous. This PR makes it just stop immediately after encountering a corrupted key, just like it would for other kinds of corruption. Let me know if there was actually some deeper meaning in this behavior and I should put it back.
- It had a few code paths silently discarding subiterator's status. The stress test caught a few.
- The backwards iteration code path was expecting the internal iterator's set of keys to be immutable. It's probably always true in practice at the moment, since ForwardIterator doesn't support backwards iteration, but this PR fixes it anyway. See added DBIteratorTest.ReverseToForwardBug for an example.
- Some parts of backwards iteration code path even did things like `assert(iter_->Valid())` after a seek, which is never a safe assumption.
- It used to not reset status on seek for some types of errors.
- Some simplifications and better comments.
- Some things got more complicated from the added error handling. I'm open to ideas for how to make it nicer.
* MergingIterator - check status after every operation on every subiterator, and in some places assert that valid subiterators have ok status.
* ForwardIterator - changed to the new convention, also slightly simplified.
* ForwardLevelIterator - fixed some bugs and simplified.
* LevelIterator - simplified.
* TwoLevelIterator - changed to the new convention. Also fixed a bug that would make SeekForPrev() sometimes silently ignore errors from first_level_iter_.
* BlockBasedTableIterator - minor changes.
* BlockIter - replaced `SetStatus()` with `Invalidate()` to make sure non-ok BlockIter is always invalid.
* PlainTableIterator - some seeks used to not reset status.
* CuckooTableIterator - tiny code cleanup.
* ManagedIterator - fixed some bugs.
* BaseDeltaIterator - changed to the new convention and fixed a bug.
* BlobDBIterator - seeks used to not reset status.
* KeyConvertingIterator - some small change.
Closes https://github.com/facebook/rocksdb/pull/3810
Differential Revision: D7888019
Pulled By: al13n321
fbshipit-source-id: 4aaf6d3421c545d16722a815b2fa2e7912bc851d
2018-05-17 09:44:14 +00:00
|
|
|
Check(0, 5000);
|
2017-08-09 22:49:40 +00:00
|
|
|
ASSERT_NOK(dbi->VerifyChecksum());
|
2020-02-14 16:15:24 +00:00
|
|
|
|
|
|
|
// In paranoid mode, the db cannot be opened due to the corrupted file.
|
|
|
|
ASSERT_TRUE(TryReopen().IsCorruption());
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, MissingDescriptor) {
|
2011-03-18 22:37:00 +00:00
|
|
|
Build(1000);
|
|
|
|
RepairDB();
|
|
|
|
Reopen();
|
|
|
|
Check(1000, 1000);
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, SequenceNumberRecovery) {
|
2011-03-18 22:37:00 +00:00
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "v1"));
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "v2"));
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "v3"));
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "v4"));
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "v5"));
|
|
|
|
RepairDB();
|
|
|
|
Reopen();
|
|
|
|
std::string v;
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
|
|
|
|
ASSERT_EQ("v5", v);
|
|
|
|
// Write something. If sequence number was not recovered properly,
|
|
|
|
// it will be hidden by an earlier write.
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "v6"));
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
|
|
|
|
ASSERT_EQ("v6", v);
|
|
|
|
Reopen();
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
|
|
|
|
ASSERT_EQ("v6", v);
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, CorruptedDescriptor) {
|
2011-03-18 22:37:00 +00:00
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), "foo", "hello"));
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2013-03-01 02:04:58 +00:00
|
|
|
dbi->TEST_CompactRange(0, nullptr, nullptr);
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
Corrupt(kDescriptorFile, 0, 1000);
|
|
|
|
Status s = TryReopen();
|
|
|
|
ASSERT_TRUE(!s.ok());
|
|
|
|
|
|
|
|
RepairDB();
|
|
|
|
Reopen();
|
|
|
|
std::string v;
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), "foo", &v));
|
|
|
|
ASSERT_EQ("hello", v);
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, CompactionInputError) {
|
2014-09-18 20:32:44 +00:00
|
|
|
Options options;
|
|
|
|
Reopen(&options);
|
2011-03-18 22:37:00 +00:00
|
|
|
Build(10);
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2015-07-17 19:02:52 +00:00
|
|
|
dbi->TEST_CompactRange(0, nullptr, nullptr);
|
|
|
|
dbi->TEST_CompactRange(1, nullptr, nullptr);
|
|
|
|
ASSERT_EQ(1, Property("rocksdb.num-files-at-level2"));
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
Corrupt(kTableFile, 100, 1);
|
|
|
|
Check(9, 9);
|
2017-08-09 22:49:40 +00:00
|
|
|
ASSERT_NOK(dbi->VerifyChecksum());
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
// Force compactions by writing lots of values
|
|
|
|
Build(10000);
|
|
|
|
Check(10000, 10000);
|
2017-08-09 22:49:40 +00:00
|
|
|
ASSERT_NOK(dbi->VerifyChecksum());
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, CompactionInputErrorParanoid) {
|
2011-03-18 22:37:00 +00:00
|
|
|
Options options;
|
|
|
|
options.paranoid_checks = true;
|
2014-04-11 19:48:38 +00:00
|
|
|
options.write_buffer_size = 131072;
|
|
|
|
options.max_write_buffer_number = 2;
|
2011-03-18 22:37:00 +00:00
|
|
|
Reopen(&options);
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2011-06-22 02:36:45 +00:00
|
|
|
|
2015-07-17 19:02:52 +00:00
|
|
|
// Fill levels >= 1
|
2012-06-23 02:30:03 +00:00
|
|
|
for (int level = 1; level < dbi->NumberLevels(); level++) {
|
2011-06-22 02:36:45 +00:00
|
|
|
dbi->Put(WriteOptions(), "", "begin");
|
|
|
|
dbi->Put(WriteOptions(), "~", "end");
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2015-07-17 19:02:52 +00:00
|
|
|
for (int comp_level = 0; comp_level < dbi->NumberLevels() - level;
|
|
|
|
++comp_level) {
|
|
|
|
dbi->TEST_CompactRange(comp_level, nullptr, nullptr);
|
|
|
|
}
|
2011-06-22 02:36:45 +00:00
|
|
|
}
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2014-04-24 18:13:28 +00:00
|
|
|
Reopen(&options);
|
|
|
|
|
2020-07-03 02:24:25 +00:00
|
|
|
dbi = static_cast_with_check<DBImpl>(db_);
|
2011-03-18 22:37:00 +00:00
|
|
|
Build(10);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2013-10-16 20:32:53 +00:00
|
|
|
dbi->TEST_WaitForCompact();
|
2013-10-05 05:32:05 +00:00
|
|
|
ASSERT_EQ(1, Property("rocksdb.num-files-at-level0"));
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2014-04-24 18:13:28 +00:00
|
|
|
CorruptTableFileAtLevel(0, 100, 1);
|
2011-03-18 22:37:00 +00:00
|
|
|
Check(9, 9);
|
2017-08-09 22:49:40 +00:00
|
|
|
ASSERT_NOK(dbi->VerifyChecksum());
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
// Write must eventually fail because of corrupted table
|
|
|
|
Status s;
|
|
|
|
std::string tmp1, tmp2;
|
2013-10-28 19:36:02 +00:00
|
|
|
bool failed = false;
|
2014-04-11 19:48:38 +00:00
|
|
|
for (int i = 0; i < 10000; i++) {
|
2011-03-18 22:37:00 +00:00
|
|
|
s = db_->Put(WriteOptions(), Key(i, &tmp1), Value(i, &tmp2));
|
2013-10-28 19:36:02 +00:00
|
|
|
if (!s.ok()) {
|
|
|
|
failed = true;
|
|
|
|
}
|
|
|
|
// if one write failed, every subsequent write must fail, too
|
|
|
|
ASSERT_TRUE(!failed || !s.ok()) << "write did not fail in a corrupted db";
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
ASSERT_TRUE(!s.ok()) << "write did not fail in corrupted paranoid db";
|
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, UnrelatedKeys) {
|
2011-03-18 22:37:00 +00:00
|
|
|
Build(10);
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2011-03-18 22:37:00 +00:00
|
|
|
Corrupt(kTableFile, 100, 1);
|
2017-08-09 22:49:40 +00:00
|
|
|
ASSERT_NOK(dbi->VerifyChecksum());
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
std::string tmp1, tmp2;
|
|
|
|
ASSERT_OK(db_->Put(WriteOptions(), Key(1000, &tmp1), Value(1000, &tmp2)));
|
|
|
|
std::string v;
|
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), Key(1000, &tmp1), &v));
|
|
|
|
ASSERT_EQ(Value(1000, &tmp2).ToString(), v);
|
2013-10-14 22:12:15 +00:00
|
|
|
dbi->TEST_FlushMemTable();
|
2011-03-18 22:37:00 +00:00
|
|
|
ASSERT_OK(db_->Get(ReadOptions(), Key(1000, &tmp1), &v));
|
|
|
|
ASSERT_EQ(Value(1000, &tmp2).ToString(), v);
|
|
|
|
}
|
|
|
|
|
2018-12-21 05:57:18 +00:00
|
|
|
TEST_F(CorruptionTest, RangeDeletionCorrupted) {
|
|
|
|
ASSERT_OK(
|
|
|
|
db_->DeleteRange(WriteOptions(), db_->DefaultColumnFamily(), "a", "b"));
|
|
|
|
ASSERT_OK(db_->Flush(FlushOptions()));
|
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
db_->GetLiveFilesMetaData(&metadata);
|
|
|
|
ASSERT_EQ(static_cast<size_t>(1), metadata.size());
|
|
|
|
std::string filename = dbname_ + metadata[0].name;
|
|
|
|
|
|
|
|
std::unique_ptr<RandomAccessFile> file;
|
|
|
|
ASSERT_OK(options_.env->NewRandomAccessFile(filename, &file, EnvOptions()));
|
|
|
|
std::unique_ptr<RandomAccessFileReader> file_reader(
|
Introduce a new storage specific Env API (#5761)
Summary:
The current Env API encompasses both storage/file operations, as well as OS related operations. Most of the APIs return a Status, which does not have enough metadata about an error, such as whether its retry-able or not, scope (i.e fault domain) of the error etc., that may be required in order to properly handle a storage error. The file APIs also do not provide enough control over the IO SLA, such as timeout, prioritization, hinting about placement and redundancy etc.
This PR separates out the file/storage APIs from Env into a new FileSystem class. The APIs are updated to return an IOStatus with metadata about the error, as well as to take an IOOptions structure as input in order to allow more control over the IO.
The user can set both ```options.env``` and ```options.file_system``` to specify that RocksDB should use the former for OS related operations and the latter for storage operations. Internally, a ```CompositeEnvWrapper``` has been introduced that inherits from ```Env``` and redirects individual methods to either an ```Env``` implementation or the ```FileSystem``` as appropriate. When options are sanitized during ```DB::Open```, ```options.env``` is replaced with a newly allocated ```CompositeEnvWrapper``` instance if both env and file_system have been specified. This way, the rest of the RocksDB code can continue to function as before.
This PR also ports PosixEnv to the new API by splitting it into two - PosixEnv and PosixFileSystem. PosixEnv is defined as a sub-class of CompositeEnvWrapper, and threading/time functions are overridden with Posix specific implementations in order to avoid an extra level of indirection.
The ```CompositeEnvWrapper``` translates ```IOStatus``` return code to ```Status```, and sets the severity to ```kSoftError``` if the io_status is retryable. The error handling code in RocksDB can then recover the DB automatically.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5761
Differential Revision: D18868376
Pulled By: anand1976
fbshipit-source-id: 39efe18a162ea746fabac6360ff529baba48486f
2019-12-13 22:47:08 +00:00
|
|
|
new RandomAccessFileReader(NewLegacyRandomAccessFileWrapper(file),
|
|
|
|
filename));
|
2018-12-21 05:57:18 +00:00
|
|
|
|
|
|
|
uint64_t file_size;
|
|
|
|
ASSERT_OK(options_.env->GetFileSize(filename, &file_size));
|
|
|
|
|
|
|
|
BlockHandle range_del_handle;
|
|
|
|
ASSERT_OK(FindMetaBlock(
|
|
|
|
file_reader.get(), file_size, kBlockBasedTableMagicNumber,
|
|
|
|
ImmutableCFOptions(options_), kRangeDelBlock, &range_del_handle));
|
|
|
|
|
|
|
|
ASSERT_OK(TryReopen());
|
2020-06-05 18:06:26 +00:00
|
|
|
test::CorruptFile(filename, static_cast<int>(range_del_handle.offset()), 1);
|
2020-02-14 16:15:24 +00:00
|
|
|
ASSERT_TRUE(TryReopen().IsCorruption());
|
2018-12-21 05:57:18 +00:00
|
|
|
}
|
|
|
|
|
2015-03-17 21:08:00 +00:00
|
|
|
TEST_F(CorruptionTest, FileSystemStateCorrupted) {
|
2014-03-20 21:18:29 +00:00
|
|
|
for (int iter = 0; iter < 2; ++iter) {
|
|
|
|
Options options;
|
|
|
|
options.paranoid_checks = true;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Reopen(&options);
|
|
|
|
Build(10);
|
|
|
|
ASSERT_OK(db_->Flush(FlushOptions()));
|
2020-07-03 02:24:25 +00:00
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
2014-03-20 21:18:29 +00:00
|
|
|
std::vector<LiveFileMetaData> metadata;
|
|
|
|
dbi->GetLiveFilesMetaData(&metadata);
|
2014-03-27 00:53:39 +00:00
|
|
|
ASSERT_GT(metadata.size(), size_t(0));
|
2014-03-20 21:18:29 +00:00
|
|
|
std::string filename = dbname_ + metadata[0].name;
|
|
|
|
|
|
|
|
delete db_;
|
2014-03-20 21:37:30 +00:00
|
|
|
db_ = nullptr;
|
2014-03-20 21:18:29 +00:00
|
|
|
|
|
|
|
if (iter == 0) { // corrupt file size
|
2018-11-09 19:17:34 +00:00
|
|
|
std::unique_ptr<WritableFile> file;
|
2014-03-20 21:18:29 +00:00
|
|
|
env_.NewWritableFile(filename, &file, EnvOptions());
|
|
|
|
file->Append(Slice("corrupted sst"));
|
|
|
|
file.reset();
|
2020-02-14 16:15:24 +00:00
|
|
|
Status x = TryReopen(&options);
|
|
|
|
ASSERT_TRUE(x.IsCorruption());
|
2014-03-20 21:18:29 +00:00
|
|
|
} else { // delete the file
|
|
|
|
env_.DeleteFile(filename);
|
2020-02-14 16:15:24 +00:00
|
|
|
Status x = TryReopen(&options);
|
|
|
|
ASSERT_TRUE(x.IsPathNotFound());
|
2014-03-20 21:18:29 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
DestroyDB(dbname_, options_);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-22 18:03:29 +00:00
|
|
|
static const auto& corruption_modes = {mock::MockTableFactory::kCorruptNone,
|
|
|
|
mock::MockTableFactory::kCorruptKey,
|
|
|
|
mock::MockTableFactory::kCorruptValue};
|
|
|
|
|
2020-08-06 21:24:24 +00:00
|
|
|
TEST_F(CorruptionTest, ParanoidFileChecksOnFlush) {
|
2020-07-22 18:03:29 +00:00
|
|
|
Options options;
|
|
|
|
options.paranoid_file_checks = true;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Status s;
|
|
|
|
for (const auto& mode : corruption_modes) {
|
|
|
|
delete db_;
|
2020-08-13 02:28:28 +00:00
|
|
|
db_ = nullptr;
|
2020-07-22 18:03:29 +00:00
|
|
|
s = DestroyDB(dbname_, options);
|
|
|
|
std::shared_ptr<mock::MockTableFactory> mock =
|
|
|
|
std::make_shared<mock::MockTableFactory>();
|
|
|
|
options.table_factory = mock;
|
|
|
|
mock->SetCorruptionMode(mode);
|
|
|
|
ASSERT_OK(DB::Open(options, dbname_, &db_));
|
2020-08-13 02:28:28 +00:00
|
|
|
assert(db_ != nullptr);
|
2020-07-22 18:03:29 +00:00
|
|
|
Build(10);
|
|
|
|
s = db_->Flush(FlushOptions());
|
|
|
|
if (mode == mock::MockTableFactory::kCorruptNone) {
|
|
|
|
ASSERT_OK(s);
|
|
|
|
} else {
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-08-06 21:24:24 +00:00
|
|
|
TEST_F(CorruptionTest, ParanoidFileChecksOnCompact) {
|
2020-07-22 18:03:29 +00:00
|
|
|
Options options;
|
|
|
|
options.paranoid_file_checks = true;
|
|
|
|
options.create_if_missing = true;
|
|
|
|
Status s;
|
|
|
|
for (const auto& mode : corruption_modes) {
|
|
|
|
delete db_;
|
2020-08-13 02:28:28 +00:00
|
|
|
db_ = nullptr;
|
2020-07-22 18:03:29 +00:00
|
|
|
s = DestroyDB(dbname_, options);
|
|
|
|
std::shared_ptr<mock::MockTableFactory> mock =
|
|
|
|
std::make_shared<mock::MockTableFactory>();
|
|
|
|
options.table_factory = mock;
|
|
|
|
ASSERT_OK(DB::Open(options, dbname_, &db_));
|
2020-08-13 02:28:28 +00:00
|
|
|
assert(db_ != nullptr);
|
2020-07-22 18:03:29 +00:00
|
|
|
Build(100, 2);
|
|
|
|
// ASSERT_OK(db_->Flush(FlushOptions()));
|
|
|
|
DBImpl* dbi = static_cast_with_check<DBImpl>(db_);
|
|
|
|
ASSERT_OK(dbi->TEST_FlushMemTable());
|
|
|
|
mock->SetCorruptionMode(mode);
|
|
|
|
s = dbi->TEST_CompactRange(0, nullptr, nullptr, nullptr, true);
|
|
|
|
if (mode == mock::MockTableFactory::kCorruptNone) {
|
|
|
|
ASSERT_OK(s);
|
|
|
|
} else {
|
|
|
|
ASSERT_NOK(s);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
2015-03-17 21:08:00 +00:00
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
2015-09-14 23:44:35 +00:00
|
|
|
|
|
|
|
#else
|
|
|
|
#include <stdio.h>
|
|
|
|
|
2018-04-16 00:19:57 +00:00
|
|
|
int main(int /*argc*/, char** /*argv*/) {
|
2015-09-14 23:44:35 +00:00
|
|
|
fprintf(stderr, "SKIPPED as RepairDB() is not supported in ROCKSDB_LITE\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif // !ROCKSDB_LITE
|