mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-26 16:30:56 +00:00
Optionally construct Post Processing Info map in MemTableInserter
Summary: MemTableInserter default constructs Post processing info std::map. However, on Windows with 2015 STL the default constructed map still dynamically allocates one node which shows up on a profiler and we loose ~40% throughput on fillrandom benchmark. Solution: declare a map as std::aligned storage and optionally construct. This addresses https://github.com/facebook/rocksdb/issues/1976 Before: ------------------------------------------------------------------- Initializing RocksDB Options from command-line flags DB path: [k:\data\BulkLoadRandom_10M_fillonly] fillrandom : 2.775 micros/op 360334 ops/sec; 280.4 MB/s Microseconds per write: Count: 10000000 Average: 2.7749 StdDev: 39.92 Min: 1 Median: 2.0826 Max: 26051 Percentiles: P50: 2.08 P75: 2.55 P99: 3.55 P99.9: 9.58 P99.99: 51.5**6 ------------------------------------------------------ After: Initializing RocksDB Options from command-line flags DB path: [k:\data\BulkLoadRandom_10M_fillon Closes https://github.com/facebook/rocksdb/pull/2011 Differential Revision: D4740823 Pulled By: siying fbshipit-source-id: 1daaa2c
This commit is contained in:
parent
e474df9470
commit
be723a8d8c
|
@ -34,6 +34,7 @@
|
||||||
#include <map>
|
#include <map>
|
||||||
#include <stack>
|
#include <stack>
|
||||||
#include <stdexcept>
|
#include <stdexcept>
|
||||||
|
#include <type_traits>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
#include "db/column_family.h"
|
#include "db/column_family.h"
|
||||||
|
@ -764,7 +765,7 @@ Status WriteBatch::RollbackToSavePoint() {
|
||||||
}
|
}
|
||||||
|
|
||||||
class MemTableInserter : public WriteBatch::Handler {
|
class MemTableInserter : public WriteBatch::Handler {
|
||||||
public:
|
|
||||||
SequenceNumber sequence_;
|
SequenceNumber sequence_;
|
||||||
ColumnFamilyMemTables* const cf_mems_;
|
ColumnFamilyMemTables* const cf_mems_;
|
||||||
FlushScheduler* const flush_scheduler_;
|
FlushScheduler* const flush_scheduler_;
|
||||||
|
@ -774,12 +775,32 @@ class MemTableInserter : public WriteBatch::Handler {
|
||||||
uint64_t log_number_ref_;
|
uint64_t log_number_ref_;
|
||||||
DBImpl* db_;
|
DBImpl* db_;
|
||||||
const bool concurrent_memtable_writes_;
|
const bool concurrent_memtable_writes_;
|
||||||
|
bool post_info_created_;
|
||||||
|
|
||||||
bool* has_valid_writes_;
|
bool* has_valid_writes_;
|
||||||
typedef std::map<MemTable*, MemTablePostProcessInfo> MemPostInfoMap;
|
// On some (!) platforms just default creating
|
||||||
MemPostInfoMap mem_post_info_map_;
|
// a map is too expensive in the Write() path as they
|
||||||
|
// cause memory allocations though unused.
|
||||||
|
// Make creation optional but do not incur
|
||||||
|
// unique_ptr additional allocation
|
||||||
|
using
|
||||||
|
MemPostInfoMap = std::map<MemTable*, MemTablePostProcessInfo>;
|
||||||
|
using
|
||||||
|
PostMapType = std::aligned_storage<sizeof(MemPostInfoMap)>::type;
|
||||||
|
PostMapType mem_post_info_map_;
|
||||||
// current recovered transaction we are rebuilding (recovery)
|
// current recovered transaction we are rebuilding (recovery)
|
||||||
WriteBatch* rebuilding_trx_;
|
WriteBatch* rebuilding_trx_;
|
||||||
|
|
||||||
|
MemPostInfoMap& GetPostMap() {
|
||||||
|
assert(concurrent_memtable_writes_);
|
||||||
|
if(!post_info_created_) {
|
||||||
|
new (&mem_post_info_map_) MemPostInfoMap();
|
||||||
|
post_info_created_ = true;
|
||||||
|
}
|
||||||
|
return *reinterpret_cast<MemPostInfoMap*>(&mem_post_info_map_);
|
||||||
|
}
|
||||||
|
|
||||||
|
public:
|
||||||
// cf_mems should not be shared with concurrent inserters
|
// cf_mems should not be shared with concurrent inserters
|
||||||
MemTableInserter(SequenceNumber sequence, ColumnFamilyMemTables* cf_mems,
|
MemTableInserter(SequenceNumber sequence, ColumnFamilyMemTables* cf_mems,
|
||||||
FlushScheduler* flush_scheduler,
|
FlushScheduler* flush_scheduler,
|
||||||
|
@ -795,18 +816,34 @@ class MemTableInserter : public WriteBatch::Handler {
|
||||||
log_number_ref_(0),
|
log_number_ref_(0),
|
||||||
db_(reinterpret_cast<DBImpl*>(db)),
|
db_(reinterpret_cast<DBImpl*>(db)),
|
||||||
concurrent_memtable_writes_(concurrent_memtable_writes),
|
concurrent_memtable_writes_(concurrent_memtable_writes),
|
||||||
|
post_info_created_(false),
|
||||||
has_valid_writes_(has_valid_writes),
|
has_valid_writes_(has_valid_writes),
|
||||||
rebuilding_trx_(nullptr) {
|
rebuilding_trx_(nullptr) {
|
||||||
assert(cf_mems_);
|
assert(cf_mems_);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
~MemTableInserter() {
|
||||||
|
if (post_info_created_) {
|
||||||
|
reinterpret_cast<MemPostInfoMap*>
|
||||||
|
(&mem_post_info_map_)->~MemPostInfoMap();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
MemTableInserter(const MemTableInserter&) = delete;
|
||||||
|
MemTableInserter& operator=(const MemTableInserter&) = delete;
|
||||||
|
|
||||||
void set_log_number_ref(uint64_t log) { log_number_ref_ = log; }
|
void set_log_number_ref(uint64_t log) { log_number_ref_ = log; }
|
||||||
|
|
||||||
SequenceNumber get_final_sequence() { return sequence_; }
|
SequenceNumber get_final_sequence() const { return sequence_; }
|
||||||
|
|
||||||
void PostProcess() {
|
void PostProcess() {
|
||||||
for (auto& pair : mem_post_info_map_) {
|
assert(concurrent_memtable_writes_);
|
||||||
pair.first->BatchPostProcess(pair.second);
|
// If post info was not created there is nothing
|
||||||
|
// to process and no need to create on demand
|
||||||
|
if(post_info_created_) {
|
||||||
|
for (auto& pair : GetPostMap()) {
|
||||||
|
pair.first->BatchPostProcess(pair.second);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1194,7 +1231,7 @@ class MemTableInserter : public WriteBatch::Handler {
|
||||||
// No need to batch counters locally if we don't use concurrent mode.
|
// No need to batch counters locally if we don't use concurrent mode.
|
||||||
return nullptr;
|
return nullptr;
|
||||||
}
|
}
|
||||||
return &mem_post_info_map_[mem];
|
return &GetPostMap()[mem];
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue