mirror of
https://github.com/facebook/rocksdb.git
synced 2024-12-02 10:15:54 +00:00
151242ce46
Summary: **Context:** The existing stat rocksdb.sst.read.micros does not reflect each of compaction and flush cases but aggregate them, which is not so helpful for us to understand IO read behavior of each of them. **Summary** - Update `StopWatch` and `RandomAccessFileReader` to record `rocksdb.sst.read.micros` and `rocksdb.file.{flush/compaction}.read.micros` - Fixed the default histogram in `RandomAccessFileReader` - New field `ReadOptions/IOOptions::io_activity`; Pass `ReadOptions` through paths under db open, flush and compaction to where we can prepare `IOOptions` and pass it to `RandomAccessFileReader` - Use `thread_status_util` for assertion in `DbStressFSWrapper` for continuous testing on we are passing correct `io_activity` under db open, flush and compaction Pull Request resolved: https://github.com/facebook/rocksdb/pull/11288 Test Plan: - **Stress test** - **Db bench 1: rocksdb.sst.read.micros COUNT ≈ sum of rocksdb.file.read.flush.micros's and rocksdb.file.read.compaction.micros's.** (without blob) - May not be exactly the same due to `HistogramStat::Add` only guarantees atomic not accuracy across threads. ``` ./db_bench -db=/dev/shm/testdb/ -statistics=true -benchmarks="fillseq" -key_size=32 -value_size=512 -num=50000 -write_buffer_size=655 -target_file_size_base=655 -disable_auto_compactions=false -compression_type=none -bloom_bits=3 (-use_plain_table=1 -prefix_size=10) ``` ``` // BlockBasedTable rocksdb.sst.read.micros P50 : 2.009374 P95 : 4.968548 P99 : 8.110362 P100 : 43.000000 COUNT : 40456 SUM : 114805 rocksdb.file.read.flush.micros P50 : 1.871841 P95 : 3.872407 P99 : 5.540541 P100 : 43.000000 COUNT : 2250 SUM : 6116 rocksdb.file.read.compaction.micros P50 : 2.023109 P95 : 5.029149 P99 : 8.196910 P100 : 26.000000 COUNT : 38206 SUM : 108689 // PlainTable Does not apply ``` - **Db bench 2: performance** **Read** SETUP: db with 900 files ``` ./db_bench -db=/dev/shm/testdb/ -benchmarks="fillseq" -key_size=32 -value_size=512 -num=50000 -write_buffer_size=655 -disable_auto_compactions=true -target_file_size_base=655 -compression_type=none ```run till convergence ``` ./db_bench -seed=1678564177044286 -use_existing_db=true -db=/dev/shm/testdb -benchmarks=readrandom[-X60] -statistics=true -num=1000000 -disable_auto_compactions=true -compression_type=none -bloom_bits=3 ``` Pre-change `readrandom [AVG 60 runs] : 21568 (± 248) ops/sec` Post-change (no regression, -0.3%) `readrandom [AVG 60 runs] : 21486 (± 236) ops/sec` **Compaction/Flush**run till convergence ``` ./db_bench -db=/dev/shm/testdb2/ -seed=1678564177044286 -benchmarks="fillseq[-X60]" -key_size=32 -value_size=512 -num=50000 -write_buffer_size=655 -disable_auto_compactions=false -target_file_size_base=655 -compression_type=none rocksdb.sst.read.micros COUNT : 33820 rocksdb.sst.read.flush.micros COUNT : 1800 rocksdb.sst.read.compaction.micros COUNT : 32020 ``` Pre-change `fillseq [AVG 46 runs] : 1391 (± 214) ops/sec; 0.7 (± 0.1) MB/sec` Post-change (no regression, ~-0.4%) `fillseq [AVG 46 runs] : 1385 (± 216) ops/sec; 0.7 (± 0.1) MB/sec` Reviewed By: ajkr Differential Revision: D44007011 Pulled By: hx235 fbshipit-source-id: a54c89e4846dfc9a135389edf3f3eedfea257132
279 lines
9.7 KiB
C++
279 lines
9.7 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
#include "db/blob/blob_file_cache.h"
|
|
|
|
#include <cassert>
|
|
#include <string>
|
|
|
|
#include "db/blob/blob_log_format.h"
|
|
#include "db/blob/blob_log_writer.h"
|
|
#include "env/mock_env.h"
|
|
#include "file/filename.h"
|
|
#include "file/read_write_util.h"
|
|
#include "file/writable_file_writer.h"
|
|
#include "options/cf_options.h"
|
|
#include "rocksdb/cache.h"
|
|
#include "rocksdb/env.h"
|
|
#include "rocksdb/file_system.h"
|
|
#include "rocksdb/options.h"
|
|
#include "rocksdb/statistics.h"
|
|
#include "test_util/sync_point.h"
|
|
#include "test_util/testharness.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
namespace {
|
|
|
|
// Creates a test blob file with a single blob in it.
|
|
void WriteBlobFile(uint32_t column_family_id,
|
|
const ImmutableOptions& immutable_options,
|
|
uint64_t blob_file_number) {
|
|
assert(!immutable_options.cf_paths.empty());
|
|
|
|
const std::string blob_file_path =
|
|
BlobFileName(immutable_options.cf_paths.front().path, blob_file_number);
|
|
|
|
std::unique_ptr<FSWritableFile> file;
|
|
ASSERT_OK(NewWritableFile(immutable_options.fs.get(), blob_file_path, &file,
|
|
FileOptions()));
|
|
|
|
std::unique_ptr<WritableFileWriter> file_writer(new WritableFileWriter(
|
|
std::move(file), blob_file_path, FileOptions(), immutable_options.clock));
|
|
|
|
constexpr Statistics* statistics = nullptr;
|
|
constexpr bool use_fsync = false;
|
|
constexpr bool do_flush = false;
|
|
|
|
BlobLogWriter blob_log_writer(std::move(file_writer), immutable_options.clock,
|
|
statistics, blob_file_number, use_fsync,
|
|
do_flush);
|
|
|
|
constexpr bool has_ttl = false;
|
|
constexpr ExpirationRange expiration_range;
|
|
|
|
BlobLogHeader header(column_family_id, kNoCompression, has_ttl,
|
|
expiration_range);
|
|
|
|
ASSERT_OK(blob_log_writer.WriteHeader(header));
|
|
|
|
constexpr char key[] = "key";
|
|
constexpr char blob[] = "blob";
|
|
|
|
std::string compressed_blob;
|
|
|
|
uint64_t key_offset = 0;
|
|
uint64_t blob_offset = 0;
|
|
|
|
ASSERT_OK(blob_log_writer.AddRecord(key, blob, &key_offset, &blob_offset));
|
|
|
|
BlobLogFooter footer;
|
|
footer.blob_count = 1;
|
|
footer.expiration_range = expiration_range;
|
|
|
|
std::string checksum_method;
|
|
std::string checksum_value;
|
|
|
|
ASSERT_OK(
|
|
blob_log_writer.AppendFooter(footer, &checksum_method, &checksum_value));
|
|
}
|
|
|
|
} // anonymous namespace
|
|
|
|
class BlobFileCacheTest : public testing::Test {
|
|
protected:
|
|
BlobFileCacheTest() { mock_env_.reset(MockEnv::Create(Env::Default())); }
|
|
|
|
std::unique_ptr<Env> mock_env_;
|
|
};
|
|
|
|
TEST_F(BlobFileCacheTest, GetBlobFileReader) {
|
|
Options options;
|
|
options.env = mock_env_.get();
|
|
options.statistics = CreateDBStatistics();
|
|
options.cf_paths.emplace_back(
|
|
test::PerThreadDBPath(mock_env_.get(),
|
|
"BlobFileCacheTest_GetBlobFileReader"),
|
|
0);
|
|
options.enable_blob_files = true;
|
|
|
|
constexpr uint32_t column_family_id = 1;
|
|
ImmutableOptions immutable_options(options);
|
|
constexpr uint64_t blob_file_number = 123;
|
|
|
|
WriteBlobFile(column_family_id, immutable_options, blob_file_number);
|
|
|
|
constexpr size_t capacity = 10;
|
|
std::shared_ptr<Cache> backing_cache = NewLRUCache(capacity);
|
|
|
|
FileOptions file_options;
|
|
constexpr HistogramImpl* blob_file_read_hist = nullptr;
|
|
|
|
BlobFileCache blob_file_cache(backing_cache.get(), &immutable_options,
|
|
&file_options, column_family_id,
|
|
blob_file_read_hist, nullptr /*IOTracer*/);
|
|
|
|
// First try: reader should be opened and put in cache
|
|
CacheHandleGuard<BlobFileReader> first;
|
|
|
|
const ReadOptions read_options;
|
|
ASSERT_OK(blob_file_cache.GetBlobFileReader(read_options, blob_file_number,
|
|
&first));
|
|
ASSERT_NE(first.GetValue(), nullptr);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_OPENS), 1);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_ERRORS), 0);
|
|
|
|
// Second try: reader should be served from cache
|
|
CacheHandleGuard<BlobFileReader> second;
|
|
|
|
ASSERT_OK(blob_file_cache.GetBlobFileReader(read_options, blob_file_number,
|
|
&second));
|
|
ASSERT_NE(second.GetValue(), nullptr);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_OPENS), 1);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_ERRORS), 0);
|
|
|
|
ASSERT_EQ(first.GetValue(), second.GetValue());
|
|
}
|
|
|
|
TEST_F(BlobFileCacheTest, GetBlobFileReader_Race) {
|
|
Options options;
|
|
options.env = mock_env_.get();
|
|
options.statistics = CreateDBStatistics();
|
|
options.cf_paths.emplace_back(
|
|
test::PerThreadDBPath(mock_env_.get(),
|
|
"BlobFileCacheTest_GetBlobFileReader_Race"),
|
|
0);
|
|
options.enable_blob_files = true;
|
|
|
|
constexpr uint32_t column_family_id = 1;
|
|
ImmutableOptions immutable_options(options);
|
|
constexpr uint64_t blob_file_number = 123;
|
|
|
|
WriteBlobFile(column_family_id, immutable_options, blob_file_number);
|
|
|
|
constexpr size_t capacity = 10;
|
|
std::shared_ptr<Cache> backing_cache = NewLRUCache(capacity);
|
|
|
|
FileOptions file_options;
|
|
constexpr HistogramImpl* blob_file_read_hist = nullptr;
|
|
|
|
BlobFileCache blob_file_cache(backing_cache.get(), &immutable_options,
|
|
&file_options, column_family_id,
|
|
blob_file_read_hist, nullptr /*IOTracer*/);
|
|
|
|
CacheHandleGuard<BlobFileReader> first;
|
|
CacheHandleGuard<BlobFileReader> second;
|
|
|
|
const ReadOptions read_options;
|
|
SyncPoint::GetInstance()->SetCallBack(
|
|
"BlobFileCache::GetBlobFileReader:DoubleCheck", [&](void* /* arg */) {
|
|
// Disabling sync points to prevent infinite recursion
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
ASSERT_OK(blob_file_cache.GetBlobFileReader(read_options,
|
|
blob_file_number, &second));
|
|
ASSERT_NE(second.GetValue(), nullptr);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_OPENS), 1);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_ERRORS), 0);
|
|
});
|
|
SyncPoint::GetInstance()->EnableProcessing();
|
|
|
|
ASSERT_OK(blob_file_cache.GetBlobFileReader(read_options, blob_file_number,
|
|
&first));
|
|
ASSERT_NE(first.GetValue(), nullptr);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_OPENS), 1);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_ERRORS), 0);
|
|
|
|
ASSERT_EQ(first.GetValue(), second.GetValue());
|
|
|
|
SyncPoint::GetInstance()->DisableProcessing();
|
|
SyncPoint::GetInstance()->ClearAllCallBacks();
|
|
}
|
|
|
|
TEST_F(BlobFileCacheTest, GetBlobFileReader_IOError) {
|
|
Options options;
|
|
options.env = mock_env_.get();
|
|
options.statistics = CreateDBStatistics();
|
|
options.cf_paths.emplace_back(
|
|
test::PerThreadDBPath(mock_env_.get(),
|
|
"BlobFileCacheTest_GetBlobFileReader_IOError"),
|
|
0);
|
|
options.enable_blob_files = true;
|
|
|
|
constexpr size_t capacity = 10;
|
|
std::shared_ptr<Cache> backing_cache = NewLRUCache(capacity);
|
|
|
|
ImmutableOptions immutable_options(options);
|
|
FileOptions file_options;
|
|
constexpr uint32_t column_family_id = 1;
|
|
constexpr HistogramImpl* blob_file_read_hist = nullptr;
|
|
|
|
BlobFileCache blob_file_cache(backing_cache.get(), &immutable_options,
|
|
&file_options, column_family_id,
|
|
blob_file_read_hist, nullptr /*IOTracer*/);
|
|
|
|
// Note: there is no blob file with the below number
|
|
constexpr uint64_t blob_file_number = 123;
|
|
|
|
CacheHandleGuard<BlobFileReader> reader;
|
|
|
|
const ReadOptions read_options;
|
|
ASSERT_TRUE(
|
|
blob_file_cache.GetBlobFileReader(read_options, blob_file_number, &reader)
|
|
.IsIOError());
|
|
ASSERT_EQ(reader.GetValue(), nullptr);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_OPENS), 1);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_ERRORS), 1);
|
|
}
|
|
|
|
TEST_F(BlobFileCacheTest, GetBlobFileReader_CacheFull) {
|
|
Options options;
|
|
options.env = mock_env_.get();
|
|
options.statistics = CreateDBStatistics();
|
|
options.cf_paths.emplace_back(
|
|
test::PerThreadDBPath(mock_env_.get(),
|
|
"BlobFileCacheTest_GetBlobFileReader_CacheFull"),
|
|
0);
|
|
options.enable_blob_files = true;
|
|
|
|
constexpr uint32_t column_family_id = 1;
|
|
ImmutableOptions immutable_options(options);
|
|
constexpr uint64_t blob_file_number = 123;
|
|
|
|
WriteBlobFile(column_family_id, immutable_options, blob_file_number);
|
|
|
|
constexpr size_t capacity = 0;
|
|
constexpr int num_shard_bits = -1; // determined automatically
|
|
constexpr bool strict_capacity_limit = true;
|
|
std::shared_ptr<Cache> backing_cache =
|
|
NewLRUCache(capacity, num_shard_bits, strict_capacity_limit);
|
|
|
|
FileOptions file_options;
|
|
constexpr HistogramImpl* blob_file_read_hist = nullptr;
|
|
|
|
BlobFileCache blob_file_cache(backing_cache.get(), &immutable_options,
|
|
&file_options, column_family_id,
|
|
blob_file_read_hist, nullptr /*IOTracer*/);
|
|
|
|
// Insert into cache should fail since it has zero capacity and
|
|
// strict_capacity_limit is set
|
|
CacheHandleGuard<BlobFileReader> reader;
|
|
|
|
const ReadOptions read_options;
|
|
ASSERT_TRUE(
|
|
blob_file_cache.GetBlobFileReader(read_options, blob_file_number, &reader)
|
|
.IsMemoryLimit());
|
|
ASSERT_EQ(reader.GetValue(), nullptr);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_OPENS), 1);
|
|
ASSERT_EQ(options.statistics->getTickerCount(NO_FILE_ERRORS), 1);
|
|
}
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
|
|
|
int main(int argc, char** argv) {
|
|
ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
return RUN_ALL_TESTS();
|
|
}
|