2016-02-09 23:12:00 +00:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2017-07-15 23:03:42 +00:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2013-10-16 21:59:46 +00:00
|
|
|
//
|
2014-03-05 18:32:54 +00:00
|
|
|
|
|
|
|
#include <sstream>
|
2022-10-25 03:45:54 +00:00
|
|
|
|
2017-04-06 02:02:00 +00:00
|
|
|
#include "monitoring/perf_context_imp.h"
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 06:09:15 +00:00
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 06:09:15 +00:00
|
|
|
|
2023-04-15 18:09:44 +00:00
|
|
|
/*
|
|
|
|
* Please add new metrics to this macro and appropriate fields will be copied,
|
|
|
|
* and/or emitted when converted to string.
|
|
|
|
* When people need to add new metrics please add the metric to the macro below
|
|
|
|
* and enclose the name of the specific metric within defCmd().
|
|
|
|
* The position of the field will be dictated by the
|
|
|
|
* order in which the macros are enumerated and the offsets of the fields will
|
|
|
|
* be matched against ''PerfContextByLevelBase'' declared in perf_context.h.
|
|
|
|
*/
|
|
|
|
// clang-format off
|
|
|
|
#define DEF_PERF_CONTEXT_LEVEL_METRICS(defCmd) \
|
|
|
|
defCmd(bloom_filter_useful) \
|
|
|
|
defCmd(bloom_filter_full_positive) \
|
|
|
|
defCmd(bloom_filter_full_true_positive) \
|
|
|
|
defCmd(user_key_return_count) \
|
|
|
|
defCmd(get_from_table_nanos) \
|
|
|
|
defCmd(block_cache_hit_count) \
|
|
|
|
defCmd(block_cache_miss_count)
|
|
|
|
// clang-format on
|
|
|
|
|
|
|
|
// Break down performance counters by level and store per-level perf context in
|
|
|
|
// PerfContextByLevel
|
|
|
|
struct PerfContextByLevelInt {
|
|
|
|
#define EMIT_FIELDS(x) uint64_t x = 0;
|
|
|
|
DEF_PERF_CONTEXT_LEVEL_METRICS(EMIT_FIELDS)
|
|
|
|
#undef EMIT_FIELDS
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Please add new metrics to this macro and appropriate fields will be copied,
|
|
|
|
* and/or emitted when converted to string.
|
|
|
|
* When people need to add new metrics please enclose the name of the specific
|
|
|
|
* metric within defCmd(). The position of the field will be dictated by the
|
|
|
|
* order in which the macros are enumerated and the offsets of the fields will
|
|
|
|
* be matched against ''PerfContextBase'' declared in perf_context.h.
|
|
|
|
*/
|
|
|
|
|
|
|
|
// clang-format off
|
|
|
|
#define DEF_PERF_CONTEXT_METRICS(defCmd) \
|
|
|
|
defCmd(user_key_comparison_count) \
|
|
|
|
defCmd(block_cache_hit_count) \
|
|
|
|
defCmd(block_read_count) \
|
|
|
|
defCmd(block_read_byte) \
|
|
|
|
defCmd(block_read_time) \
|
|
|
|
defCmd(block_read_cpu_time) \
|
|
|
|
defCmd(block_cache_index_hit_count) \
|
|
|
|
defCmd(block_cache_standalone_handle_count) \
|
|
|
|
defCmd(block_cache_real_handle_count) \
|
|
|
|
defCmd(index_block_read_count) \
|
|
|
|
defCmd(block_cache_filter_hit_count) \
|
|
|
|
defCmd(filter_block_read_count) \
|
|
|
|
defCmd(compression_dict_block_read_count) \
|
2024-03-21 17:46:46 +00:00
|
|
|
defCmd(block_cache_index_read_byte) \
|
|
|
|
defCmd(block_cache_filter_read_byte) \
|
|
|
|
defCmd(block_cache_compression_dict_read_byte) \
|
|
|
|
defCmd(block_cache_read_byte) \
|
2023-04-15 18:09:44 +00:00
|
|
|
defCmd(secondary_cache_hit_count) \
|
|
|
|
defCmd(compressed_sec_cache_insert_real_count) \
|
|
|
|
defCmd(compressed_sec_cache_insert_dummy_count) \
|
|
|
|
defCmd(compressed_sec_cache_uncompressed_bytes) \
|
|
|
|
defCmd(compressed_sec_cache_compressed_bytes) \
|
|
|
|
defCmd(block_checksum_time) \
|
|
|
|
defCmd(block_decompress_time) \
|
|
|
|
defCmd(get_read_bytes) \
|
|
|
|
defCmd(multiget_read_bytes) \
|
|
|
|
defCmd(iter_read_bytes) \
|
|
|
|
defCmd(blob_cache_hit_count) \
|
|
|
|
defCmd(blob_read_count) \
|
|
|
|
defCmd(blob_read_byte) \
|
|
|
|
defCmd(blob_read_time) \
|
|
|
|
defCmd(blob_checksum_time) \
|
|
|
|
defCmd(blob_decompress_time) \
|
|
|
|
defCmd(internal_key_skipped_count) \
|
|
|
|
defCmd(internal_delete_skipped_count) \
|
|
|
|
defCmd(internal_recent_skipped_count) \
|
|
|
|
defCmd(internal_merge_count) \
|
|
|
|
defCmd(internal_merge_point_lookup_count) \
|
|
|
|
defCmd(internal_range_del_reseek_count) \
|
|
|
|
defCmd(get_snapshot_time) \
|
|
|
|
defCmd(get_from_memtable_time) \
|
|
|
|
defCmd(get_from_memtable_count) \
|
|
|
|
defCmd(get_post_process_time) \
|
|
|
|
defCmd(get_from_output_files_time) \
|
|
|
|
defCmd(seek_on_memtable_time) \
|
|
|
|
defCmd(seek_on_memtable_count) \
|
|
|
|
defCmd(next_on_memtable_count) \
|
|
|
|
defCmd(prev_on_memtable_count) \
|
|
|
|
defCmd(seek_child_seek_time) \
|
|
|
|
defCmd(seek_child_seek_count) \
|
|
|
|
defCmd(seek_min_heap_time) \
|
|
|
|
defCmd(seek_max_heap_time) \
|
|
|
|
defCmd(seek_internal_seek_time) \
|
|
|
|
defCmd(find_next_user_entry_time) \
|
|
|
|
defCmd(write_wal_time) \
|
|
|
|
defCmd(write_memtable_time) \
|
|
|
|
defCmd(write_delay_time) \
|
|
|
|
defCmd(write_scheduling_flushes_compactions_time)\
|
|
|
|
defCmd(write_pre_and_post_process_time) \
|
|
|
|
defCmd(write_thread_wait_nanos) \
|
|
|
|
defCmd(db_mutex_lock_nanos) \
|
|
|
|
defCmd(db_condition_wait_nanos) \
|
|
|
|
defCmd(merge_operator_time_nanos) \
|
|
|
|
defCmd(read_index_block_nanos) \
|
|
|
|
defCmd(read_filter_block_nanos) \
|
|
|
|
defCmd(new_table_block_iter_nanos) \
|
|
|
|
defCmd(new_table_iterator_nanos) \
|
|
|
|
defCmd(block_seek_nanos) \
|
|
|
|
defCmd(find_table_nanos) \
|
|
|
|
defCmd(bloom_memtable_hit_count) \
|
|
|
|
defCmd(bloom_memtable_miss_count) \
|
|
|
|
defCmd(bloom_sst_hit_count) \
|
|
|
|
defCmd(bloom_sst_miss_count) \
|
|
|
|
defCmd(key_lock_wait_time) \
|
|
|
|
defCmd(key_lock_wait_count) \
|
|
|
|
defCmd(env_new_sequential_file_nanos) \
|
|
|
|
defCmd(env_new_random_access_file_nanos) \
|
|
|
|
defCmd(env_new_writable_file_nanos) \
|
|
|
|
defCmd(env_reuse_writable_file_nanos) \
|
|
|
|
defCmd(env_new_random_rw_file_nanos) \
|
|
|
|
defCmd(env_new_directory_nanos) \
|
|
|
|
defCmd(env_file_exists_nanos) \
|
|
|
|
defCmd(env_get_children_nanos) \
|
|
|
|
defCmd(env_get_children_file_attributes_nanos) \
|
|
|
|
defCmd(env_delete_file_nanos) \
|
|
|
|
defCmd(env_create_dir_nanos) \
|
|
|
|
defCmd(env_create_dir_if_missing_nanos) \
|
|
|
|
defCmd(env_delete_dir_nanos) \
|
|
|
|
defCmd(env_get_file_size_nanos) \
|
|
|
|
defCmd(env_get_file_modification_time_nanos) \
|
|
|
|
defCmd(env_rename_file_nanos) \
|
|
|
|
defCmd(env_link_file_nanos) \
|
|
|
|
defCmd(env_lock_file_nanos) \
|
|
|
|
defCmd(env_unlock_file_nanos) \
|
|
|
|
defCmd(env_new_logger_nanos) \
|
|
|
|
defCmd(get_cpu_nanos) \
|
|
|
|
defCmd(iter_next_cpu_nanos) \
|
|
|
|
defCmd(iter_prev_cpu_nanos) \
|
|
|
|
defCmd(iter_seek_cpu_nanos) \
|
|
|
|
defCmd(iter_next_count) \
|
|
|
|
defCmd(iter_prev_count) \
|
|
|
|
defCmd(iter_seek_count) \
|
|
|
|
defCmd(encrypt_data_nanos) \
|
|
|
|
defCmd(decrypt_data_nanos) \
|
|
|
|
defCmd(number_async_seek)
|
|
|
|
// clang-format on
|
|
|
|
|
|
|
|
struct PerfContextInt {
|
|
|
|
#define EMIT_FIELDS(x) uint64_t x;
|
|
|
|
DEF_PERF_CONTEXT_METRICS(EMIT_FIELDS)
|
|
|
|
#undef EMIT_FIELDS
|
|
|
|
};
|
|
|
|
|
2021-04-13 14:55:58 +00:00
|
|
|
#if defined(NPERF_CONTEXT)
|
|
|
|
// Should not be used because the counters are not thread-safe.
|
|
|
|
// Put here just to make get_perf_context() simple without ifdef.
|
2017-06-03 00:12:39 +00:00
|
|
|
PerfContext perf_context;
|
2021-04-13 14:55:58 +00:00
|
|
|
#else
|
2022-05-18 22:25:19 +00:00
|
|
|
thread_local PerfContext perf_context;
|
2014-04-04 20:11:44 +00:00
|
|
|
#endif
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 06:09:15 +00:00
|
|
|
|
2023-04-15 18:09:44 +00:00
|
|
|
PerfContext* get_perf_context() {
|
|
|
|
static_assert(sizeof(PerfContextBase) == sizeof(PerfContextInt));
|
|
|
|
static_assert(sizeof(PerfContextByLevelBase) ==
|
|
|
|
sizeof(PerfContextByLevelInt));
|
|
|
|
/*
|
|
|
|
* Validate that we have the same fields and offsets between the external user
|
|
|
|
* facing
|
|
|
|
* ''PerfContextBase'' and ''PerfContextByLevelBase' structures with the
|
|
|
|
* internal structures that we generate from the DEF_* macros above. This way
|
|
|
|
* if people add metrics to the user-facing header file, they will have a
|
|
|
|
* build failure and need to add similar fields to the macros in this file.
|
|
|
|
* These are compile-time validations and don't impose any run-time penalties.
|
|
|
|
*/
|
|
|
|
#define EMIT_OFFSET_ASSERTION(x) \
|
|
|
|
static_assert(offsetof(PerfContextBase, x) == offsetof(PerfContextInt, x));
|
|
|
|
DEF_PERF_CONTEXT_METRICS(EMIT_OFFSET_ASSERTION)
|
|
|
|
#undef EMIT_OFFSET_ASSERTION
|
|
|
|
#define EMIT_OFFSET_ASSERTION(x) \
|
|
|
|
static_assert(offsetof(PerfContextByLevelBase, x) == \
|
|
|
|
offsetof(PerfContextByLevelInt, x));
|
|
|
|
DEF_PERF_CONTEXT_LEVEL_METRICS(EMIT_OFFSET_ASSERTION)
|
|
|
|
#undef EMIT_OFFSET_ASSERTION
|
|
|
|
return &perf_context;
|
|
|
|
}
|
2017-06-03 00:12:39 +00:00
|
|
|
|
2018-10-17 18:18:00 +00:00
|
|
|
PerfContext::~PerfContext() {
|
2022-05-18 22:25:19 +00:00
|
|
|
#if !defined(NPERF_CONTEXT) && !defined(OS_SOLARIS)
|
2018-10-17 18:18:00 +00:00
|
|
|
ClearPerLevelPerfContext();
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2019-02-05 22:12:08 +00:00
|
|
|
PerfContext::PerfContext(const PerfContext& other) {
|
2020-04-03 20:21:53 +00:00
|
|
|
#ifdef NPERF_CONTEXT
|
|
|
|
(void)other;
|
|
|
|
#else
|
2023-04-15 18:09:44 +00:00
|
|
|
copyMetrics(&other);
|
2019-02-05 22:12:08 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
PerfContext::PerfContext(PerfContext&& other) noexcept {
|
2020-04-03 20:21:53 +00:00
|
|
|
#ifdef NPERF_CONTEXT
|
|
|
|
(void)other;
|
|
|
|
#else
|
2023-04-15 18:09:44 +00:00
|
|
|
copyMetrics(&other);
|
2019-02-05 22:12:08 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
PerfContext& PerfContext::operator=(const PerfContext& other) {
|
2020-04-03 20:21:53 +00:00
|
|
|
#ifdef NPERF_CONTEXT
|
|
|
|
(void)other;
|
|
|
|
#else
|
2023-04-15 18:09:44 +00:00
|
|
|
copyMetrics(&other);
|
|
|
|
#endif
|
|
|
|
return *this;
|
|
|
|
}
|
2019-02-05 22:12:08 +00:00
|
|
|
|
2023-04-15 18:09:44 +00:00
|
|
|
void PerfContext::copyMetrics(const PerfContext* other) noexcept {
|
|
|
|
#ifdef NPERF_CONTEXT
|
|
|
|
(void)other;
|
|
|
|
#else
|
|
|
|
#define EMIT_COPY_FIELDS(x) x = other->x;
|
|
|
|
DEF_PERF_CONTEXT_METRICS(EMIT_COPY_FIELDS)
|
|
|
|
#undef EMIT_COPY_FIELDS
|
2019-02-05 22:12:08 +00:00
|
|
|
if (per_level_perf_context_enabled && level_to_perf_context != nullptr) {
|
|
|
|
ClearPerLevelPerfContext();
|
|
|
|
}
|
2023-04-15 18:09:44 +00:00
|
|
|
if (other->level_to_perf_context != nullptr) {
|
2019-02-05 22:12:08 +00:00
|
|
|
level_to_perf_context = new std::map<uint32_t, PerfContextByLevel>();
|
2023-04-15 18:09:44 +00:00
|
|
|
*level_to_perf_context = *other->level_to_perf_context;
|
2019-02-05 22:12:08 +00:00
|
|
|
}
|
2023-04-15 18:09:44 +00:00
|
|
|
per_level_perf_context_enabled = other->per_level_perf_context_enabled;
|
2019-02-05 22:12:08 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 06:09:15 +00:00
|
|
|
void PerfContext::Reset() {
|
2017-06-03 00:12:39 +00:00
|
|
|
#ifndef NPERF_CONTEXT
|
2023-04-15 18:09:44 +00:00
|
|
|
#define EMIT_FIELDS(x) x = 0;
|
|
|
|
DEF_PERF_CONTEXT_METRICS(EMIT_FIELDS)
|
|
|
|
#undef EMIT_FIELDS
|
2018-10-17 18:18:00 +00:00
|
|
|
if (per_level_perf_context_enabled && level_to_perf_context) {
|
|
|
|
for (auto& kv : *level_to_perf_context) {
|
|
|
|
kv.second.Reset();
|
|
|
|
}
|
|
|
|
}
|
2014-04-08 17:58:07 +00:00
|
|
|
#endif
|
[RocksDB] Added nano second stopwatch and new perf counters to track block read cost
Summary: The pupose of this diff is to expose per user-call level precise timing of block read, so that we can answer questions like: a Get() costs me 100ms, is that somehow related to loading blocks from file system, or sth else? We will answer that with EXACTLY how many blocks have been read, how much time was spent on transfering the bytes from os, how much time was spent on checksum verification and how much time was spent on block decompression, just for that one Get. A nano second stopwatch was introduced to track time with higher precision. The cost/precision of the stopwatch is also measured in unit-test. On my dev box, retrieving one time instance costs about 30ns, on average. The deviation of timing results is good enough to track 100ns-1us level events. And the overhead could be safely ignored for 100us level events (10000 instances/s), for example, a viewstate thrift call.
Test Plan: perf_context_test, also testing with viewstate shadow traffic.
Reviewers: dhruba
Reviewed By: dhruba
CC: leveldb, xjin
Differential Revision: https://reviews.facebook.net/D12351
2013-06-04 06:09:15 +00:00
|
|
|
}
|
|
|
|
|
2018-10-17 18:18:00 +00:00
|
|
|
void PerfContextByLevel::Reset() {
|
|
|
|
#ifndef NPERF_CONTEXT
|
2023-04-15 18:09:44 +00:00
|
|
|
#define EMIT_FIELDS(x) x = 0;
|
|
|
|
DEF_PERF_CONTEXT_LEVEL_METRICS(EMIT_FIELDS)
|
|
|
|
#undef EMIT_FIELDS
|
2018-10-17 18:18:00 +00:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2016-02-01 21:41:13 +00:00
|
|
|
std::string PerfContext::ToString(bool exclude_zero_counters) const {
|
2017-06-03 00:12:39 +00:00
|
|
|
#ifdef NPERF_CONTEXT
|
2020-04-03 20:21:53 +00:00
|
|
|
(void)exclude_zero_counters;
|
2014-04-08 17:58:07 +00:00
|
|
|
return "";
|
|
|
|
#else
|
2014-03-05 18:32:54 +00:00
|
|
|
std::ostringstream ss;
|
2023-04-15 18:09:44 +00:00
|
|
|
#define PERF_CONTEXT_OUTPUT(counter) \
|
|
|
|
if (!exclude_zero_counters || (counter > 0)) { \
|
|
|
|
ss << #counter << " = " << counter << ", "; \
|
|
|
|
}
|
|
|
|
DEF_PERF_CONTEXT_METRICS(PERF_CONTEXT_OUTPUT)
|
|
|
|
#undef PERF_CONTEXT_OUTPUT
|
|
|
|
if (per_level_perf_context_enabled && level_to_perf_context) {
|
|
|
|
#define PERF_CONTEXT_BY_LEVEL_OUTPUT_ONE_COUNTER(counter) \
|
|
|
|
ss << #counter << " = "; \
|
|
|
|
for (auto& kv : *level_to_perf_context) { \
|
|
|
|
if (!exclude_zero_counters || (kv.second.counter > 0)) { \
|
|
|
|
ss << kv.second.counter << "@level" << kv.first << ", "; \
|
|
|
|
} \
|
|
|
|
}
|
|
|
|
DEF_PERF_CONTEXT_LEVEL_METRICS(PERF_CONTEXT_BY_LEVEL_OUTPUT_ONE_COUNTER)
|
|
|
|
#undef PERF_CONTEXT_BY_LEVEL_OUTPUT_ONE_COUNTER
|
|
|
|
}
|
2019-09-03 18:21:47 +00:00
|
|
|
std::string str = ss.str();
|
|
|
|
str.erase(str.find_last_not_of(", ") + 1);
|
|
|
|
return str;
|
2014-04-08 17:58:07 +00:00
|
|
|
#endif
|
2014-03-05 18:32:54 +00:00
|
|
|
}
|
|
|
|
|
2018-10-17 18:18:00 +00:00
|
|
|
void PerfContext::EnablePerLevelPerfContext() {
|
2019-02-05 22:12:08 +00:00
|
|
|
if (level_to_perf_context == nullptr) {
|
2018-10-17 18:18:00 +00:00
|
|
|
level_to_perf_context = new std::map<uint32_t, PerfContextByLevel>();
|
|
|
|
}
|
|
|
|
per_level_perf_context_enabled = true;
|
|
|
|
}
|
|
|
|
|
2022-10-25 03:45:54 +00:00
|
|
|
void PerfContext::DisablePerLevelPerfContext() {
|
2018-10-17 18:18:00 +00:00
|
|
|
per_level_perf_context_enabled = false;
|
|
|
|
}
|
|
|
|
|
2022-10-25 03:45:54 +00:00
|
|
|
void PerfContext::ClearPerLevelPerfContext() {
|
2019-02-05 22:12:08 +00:00
|
|
|
if (level_to_perf_context != nullptr) {
|
|
|
|
level_to_perf_context->clear();
|
2018-10-17 18:18:00 +00:00
|
|
|
delete level_to_perf_context;
|
|
|
|
level_to_perf_context = nullptr;
|
|
|
|
}
|
|
|
|
per_level_perf_context_enabled = false;
|
|
|
|
}
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|