mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-25 22:44:05 +00:00
d665afdbf3
Summary: **Context/Summary:** As revealed by heap profiling, allocation of `FileMetaData` for [newly created file added to a Version](https://github.com/facebook/rocksdb/pull/9924/files#diff-a6aa385940793f95a2c5b39cc670bd440c4547fa54fd44622f756382d5e47e43R774) can consume significant heap memory. This PR is to account that toward our global memory limit based on block cache capacity. Pull Request resolved: https://github.com/facebook/rocksdb/pull/9924 Test Plan: - Previous `make check` verified there are only 2 places where the memory of the allocated `FileMetaData` can be released - New unit test `TEST_P(ChargeFileMetadataTestWithParam, Basic)` - db bench (CPU cost of `charge_file_metadata` in write and compact) - **write micros/op: -0.24%** : `TEST_TMPDIR=/dev/shm/testdb ./db_bench -benchmarks=fillseq -db=$TEST_TMPDIR -charge_file_metadata=1 (remove this option for pre-PR) -disable_auto_compactions=1 -write_buffer_size=100000 -num=4000000 | egrep 'fillseq'` - **compact micros/op -0.87%** : `TEST_TMPDIR=/dev/shm/testdb ./db_bench -benchmarks=fillseq -db=$TEST_TMPDIR -charge_file_metadata=1 -disable_auto_compactions=1 -write_buffer_size=100000 -num=4000000 -numdistinct=1000 && ./db_bench -benchmarks=compact -db=$TEST_TMPDIR -use_existing_db=1 -charge_file_metadata=1 -disable_auto_compactions=1 | egrep 'compact'` table 1 - write #-run | (pre-PR) avg micros/op | std micros/op | (post-PR) micros/op | std micros/op | change (%) -- | -- | -- | -- | -- | -- 10 | 3.9711 | 0.264408 | 3.9914 | 0.254563 | 0.5111933721 20 | 3.83905 | 0.0664488 | 3.8251 | 0.0695456 | -0.3633711465 40 | 3.86625 | 0.136669 | 3.8867 | 0.143765 | 0.5289363078 80 | 3.87828 | 0.119007 | 3.86791 | 0.115674 | **-0.2673865734** 160 | 3.87677 | 0.162231 | 3.86739 | 0.16663 | **-0.2419539978** table 2 - compact #-run | (pre-PR) avg micros/op | std micros/op | (post-PR) micros/op | std micros/op | change (%) -- | -- | -- | -- | -- | -- 10 | 2,399,650.00 | 96,375.80 | 2,359,537.00 | 53,243.60 | -1.67 20 | 2,410,480.00 | 89,988.00 | 2,433,580.00 | 91,121.20 | 0.96 40 | 2.41E+06 | 121811 | 2.39E+06 | 131525 | **-0.96** 80 | 2.40E+06 | 134503 | 2.39E+06 | 108799 | **-0.78** - stress test: `python3 tools/db_crashtest.py blackbox --charge_file_metadata=1 --cache_size=1` killed as normal Reviewed By: ajkr Differential Revision: D36055583 Pulled By: hx235 fbshipit-source-id: b60eab94707103cb1322cf815f05810ef0232625
317 lines
13 KiB
C++
317 lines
13 KiB
C++
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#pragma once
|
|
|
|
#include <atomic>
|
|
#include <cstddef>
|
|
#include <cstdint>
|
|
#include <memory>
|
|
#include <mutex>
|
|
#include <vector>
|
|
|
|
#include "cache/cache_entry_roles.h"
|
|
#include "cache/cache_key.h"
|
|
#include "rocksdb/cache.h"
|
|
#include "rocksdb/slice.h"
|
|
#include "rocksdb/status.h"
|
|
#include "util/coding.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
// CacheReservationManager is an interface for reserving cache space for the
|
|
// memory used
|
|
class CacheReservationManager {
|
|
public:
|
|
// CacheReservationHandle is for managing the lifetime of a cache reservation
|
|
// for an incremental amount of memory used (i.e, incremental_memory_used)
|
|
class CacheReservationHandle {
|
|
public:
|
|
virtual ~CacheReservationHandle() {}
|
|
};
|
|
virtual ~CacheReservationManager() {}
|
|
virtual Status UpdateCacheReservation(std::size_t new_memory_used) = 0;
|
|
// TODO(hx235): replace the usage of
|
|
// `UpdateCacheReservation(memory_used_delta, increase)` with
|
|
// `UpdateCacheReservation(new_memory_used)` so that we only have one
|
|
// `UpdateCacheReservation` function
|
|
virtual Status UpdateCacheReservation(std::size_t memory_used_delta,
|
|
bool increase) = 0;
|
|
virtual Status MakeCacheReservation(
|
|
std::size_t incremental_memory_used,
|
|
std::unique_ptr<CacheReservationManager::CacheReservationHandle>
|
|
*handle) = 0;
|
|
virtual std::size_t GetTotalReservedCacheSize() = 0;
|
|
virtual std::size_t GetTotalMemoryUsed() = 0;
|
|
};
|
|
|
|
// CacheReservationManagerImpl implements interface CacheReservationManager
|
|
// for reserving cache space for the memory used by inserting/releasing dummy
|
|
// entries in the cache.
|
|
//
|
|
// This class is NOT thread-safe, except that GetTotalReservedCacheSize()
|
|
// can be called without external synchronization.
|
|
template <CacheEntryRole R>
|
|
class CacheReservationManagerImpl
|
|
: public CacheReservationManager,
|
|
public std::enable_shared_from_this<CacheReservationManagerImpl<R>> {
|
|
public:
|
|
class CacheReservationHandle
|
|
: public CacheReservationManager::CacheReservationHandle {
|
|
public:
|
|
CacheReservationHandle(
|
|
std::size_t incremental_memory_used,
|
|
std::shared_ptr<CacheReservationManagerImpl> cache_res_mgr);
|
|
~CacheReservationHandle() override;
|
|
|
|
private:
|
|
std::size_t incremental_memory_used_;
|
|
std::shared_ptr<CacheReservationManagerImpl> cache_res_mgr_;
|
|
};
|
|
|
|
// Construct a CacheReservationManagerImpl
|
|
// @param cache The cache where dummy entries are inserted and released for
|
|
// reserving cache space
|
|
// @param delayed_decrease If set true, then dummy entries won't be released
|
|
// immediately when memory usage decreases.
|
|
// Instead, it will be released when the memory usage
|
|
// decreases to 3/4 of what we have reserved so far.
|
|
// This is for saving some future dummy entry
|
|
// insertion when memory usage increases are likely to
|
|
// happen in the near future.
|
|
//
|
|
// REQUIRED: cache is not nullptr
|
|
explicit CacheReservationManagerImpl(std::shared_ptr<Cache> cache,
|
|
bool delayed_decrease = false);
|
|
|
|
// no copy constructor, copy assignment, move constructor, move assignment
|
|
CacheReservationManagerImpl(const CacheReservationManagerImpl &) = delete;
|
|
CacheReservationManagerImpl &operator=(const CacheReservationManagerImpl &) =
|
|
delete;
|
|
CacheReservationManagerImpl(CacheReservationManagerImpl &&) = delete;
|
|
CacheReservationManagerImpl &operator=(CacheReservationManagerImpl &&) =
|
|
delete;
|
|
|
|
~CacheReservationManagerImpl() override;
|
|
|
|
// One of the two ways of reserving/releasing cache space,
|
|
// see MakeCacheReservation() for the other.
|
|
//
|
|
// Use ONLY one of these two ways to prevent unexpected behavior.
|
|
//
|
|
// Insert and release dummy entries in the cache to
|
|
// match the size of total dummy entries with the least multiple of
|
|
// kSizeDummyEntry greater than or equal to new_mem_used
|
|
//
|
|
// Insert dummy entries if new_memory_used > cache_allocated_size_;
|
|
//
|
|
// Release dummy entries if new_memory_used < cache_allocated_size_
|
|
// (and new_memory_used < cache_allocated_size_ * 3/4
|
|
// when delayed_decrease is set true);
|
|
//
|
|
// Keey dummy entries the same if (1) new_memory_used == cache_allocated_size_
|
|
// or (2) new_memory_used is in the interval of
|
|
// [cache_allocated_size_ * 3/4, cache_allocated_size) when delayed_decrease
|
|
// is set true.
|
|
//
|
|
// @param new_memory_used The number of bytes used by new memory
|
|
// The most recent new_memoy_used passed in will be returned
|
|
// in GetTotalMemoryUsed() even when the call return non-ok status.
|
|
//
|
|
// Since the class is NOT thread-safe, external synchronization on the
|
|
// order of calling UpdateCacheReservation() is needed if you want
|
|
// GetTotalMemoryUsed() indeed returns the latest memory used.
|
|
//
|
|
// @return On inserting dummy entries, it returns Status::OK() if all dummy
|
|
// entry insertions succeed.
|
|
// Otherwise, it returns the first non-ok status;
|
|
// On releasing dummy entries, it always returns Status::OK().
|
|
// On keeping dummy entries the same, it always returns Status::OK().
|
|
Status UpdateCacheReservation(std::size_t new_memory_used) override;
|
|
|
|
Status UpdateCacheReservation(std::size_t /* memory_used_delta */,
|
|
bool /* increase */) override {
|
|
return Status::NotSupported();
|
|
}
|
|
|
|
// One of the two ways of reserving cache space and releasing is done through
|
|
// destruction of CacheReservationHandle.
|
|
// See UpdateCacheReservation() for the other way.
|
|
//
|
|
// Use ONLY one of these two ways to prevent unexpected behavior.
|
|
//
|
|
// Insert dummy entries in the cache for the incremental memory usage
|
|
// to match the size of total dummy entries with the least multiple of
|
|
// kSizeDummyEntry greater than or equal to the total memory used.
|
|
//
|
|
// A CacheReservationHandle is returned as an output parameter.
|
|
// The reserved dummy entries are automatically released on the destruction of
|
|
// this handle, which achieves better RAII per cache reservation.
|
|
//
|
|
// WARNING: Deallocate all the handles of the CacheReservationManager object
|
|
// before deallocating the object to prevent unexpected behavior.
|
|
//
|
|
// @param incremental_memory_used The number of bytes increased in memory
|
|
// usage.
|
|
//
|
|
// Calling GetTotalMemoryUsed() afterward will return the total memory
|
|
// increased by this number, even when calling MakeCacheReservation()
|
|
// returns non-ok status.
|
|
//
|
|
// Since the class is NOT thread-safe, external synchronization in
|
|
// calling MakeCacheReservation() is needed if you want
|
|
// GetTotalMemoryUsed() indeed returns the latest memory used.
|
|
//
|
|
// @param handle An pointer to std::unique_ptr<CacheReservationHandle> that
|
|
// manages the lifetime of the cache reservation represented by the
|
|
// handle.
|
|
//
|
|
// @return It returns Status::OK() if all dummy
|
|
// entry insertions succeed.
|
|
// Otherwise, it returns the first non-ok status;
|
|
//
|
|
// REQUIRES: handle != nullptr
|
|
Status MakeCacheReservation(
|
|
std::size_t incremental_memory_used,
|
|
std::unique_ptr<CacheReservationManager::CacheReservationHandle> *handle)
|
|
override;
|
|
|
|
// Return the size of the cache (which is a multiple of kSizeDummyEntry)
|
|
// successfully reserved by calling UpdateCacheReservation().
|
|
//
|
|
// When UpdateCacheReservation() returns non-ok status,
|
|
// calling GetTotalReservedCacheSize() after that might return a slightly
|
|
// smaller number than the actual reserved cache size due to
|
|
// the returned number will always be a multiple of kSizeDummyEntry
|
|
// and cache full might happen in the middle of inserting a dummy entry.
|
|
std::size_t GetTotalReservedCacheSize() override;
|
|
|
|
// Return the latest total memory used indicated by the most recent call of
|
|
// UpdateCacheReservation(std::size_t new_memory_used);
|
|
std::size_t GetTotalMemoryUsed() override;
|
|
|
|
static constexpr std::size_t GetDummyEntrySize() { return kSizeDummyEntry; }
|
|
|
|
// For testing only - it is to help ensure the NoopDeleterForRole<R>
|
|
// accessed from CacheReservationManagerImpl and the one accessed from the
|
|
// test are from the same translation units
|
|
static Cache::DeleterFn TEST_GetNoopDeleterForRole();
|
|
|
|
private:
|
|
static constexpr std::size_t kSizeDummyEntry = 256 * 1024;
|
|
|
|
Slice GetNextCacheKey();
|
|
|
|
Status ReleaseCacheReservation(std::size_t incremental_memory_used);
|
|
Status IncreaseCacheReservation(std::size_t new_mem_used);
|
|
Status DecreaseCacheReservation(std::size_t new_mem_used);
|
|
|
|
std::shared_ptr<Cache> cache_;
|
|
bool delayed_decrease_;
|
|
std::atomic<std::size_t> cache_allocated_size_;
|
|
std::size_t memory_used_;
|
|
std::vector<Cache::Handle *> dummy_handles_;
|
|
CacheKey cache_key_;
|
|
};
|
|
|
|
class ConcurrentCacheReservationManager
|
|
: public CacheReservationManager,
|
|
public std::enable_shared_from_this<ConcurrentCacheReservationManager> {
|
|
public:
|
|
class CacheReservationHandle
|
|
: public CacheReservationManager::CacheReservationHandle {
|
|
public:
|
|
CacheReservationHandle(
|
|
std::shared_ptr<ConcurrentCacheReservationManager> cache_res_mgr,
|
|
std::unique_ptr<CacheReservationManager::CacheReservationHandle>
|
|
cache_res_handle) {
|
|
assert(cache_res_mgr && cache_res_handle);
|
|
cache_res_mgr_ = cache_res_mgr;
|
|
cache_res_handle_ = std::move(cache_res_handle);
|
|
}
|
|
|
|
~CacheReservationHandle() override {
|
|
std::lock_guard<std::mutex> lock(cache_res_mgr_->cache_res_mgr_mu_);
|
|
cache_res_handle_.reset();
|
|
}
|
|
|
|
private:
|
|
std::shared_ptr<ConcurrentCacheReservationManager> cache_res_mgr_;
|
|
std::unique_ptr<CacheReservationManager::CacheReservationHandle>
|
|
cache_res_handle_;
|
|
};
|
|
|
|
explicit ConcurrentCacheReservationManager(
|
|
std::shared_ptr<CacheReservationManager> cache_res_mgr) {
|
|
cache_res_mgr_ = std::move(cache_res_mgr);
|
|
}
|
|
ConcurrentCacheReservationManager(const ConcurrentCacheReservationManager &) =
|
|
delete;
|
|
ConcurrentCacheReservationManager &operator=(
|
|
const ConcurrentCacheReservationManager &) = delete;
|
|
ConcurrentCacheReservationManager(ConcurrentCacheReservationManager &&) =
|
|
delete;
|
|
ConcurrentCacheReservationManager &operator=(
|
|
ConcurrentCacheReservationManager &&) = delete;
|
|
|
|
~ConcurrentCacheReservationManager() override {}
|
|
|
|
inline Status UpdateCacheReservation(std::size_t new_memory_used) override {
|
|
std::lock_guard<std::mutex> lock(cache_res_mgr_mu_);
|
|
return cache_res_mgr_->UpdateCacheReservation(new_memory_used);
|
|
}
|
|
|
|
inline Status UpdateCacheReservation(std::size_t memory_used_delta,
|
|
bool increase) override {
|
|
std::lock_guard<std::mutex> lock(cache_res_mgr_mu_);
|
|
std::size_t total_mem_used = cache_res_mgr_->GetTotalMemoryUsed();
|
|
Status s;
|
|
if (!increase) {
|
|
assert(total_mem_used >= memory_used_delta);
|
|
s = cache_res_mgr_->UpdateCacheReservation(total_mem_used -
|
|
memory_used_delta);
|
|
} else {
|
|
s = cache_res_mgr_->UpdateCacheReservation(total_mem_used +
|
|
memory_used_delta);
|
|
}
|
|
return s;
|
|
}
|
|
|
|
inline Status MakeCacheReservation(
|
|
std::size_t incremental_memory_used,
|
|
std::unique_ptr<CacheReservationManager::CacheReservationHandle> *handle)
|
|
override {
|
|
std::unique_ptr<CacheReservationManager::CacheReservationHandle>
|
|
wrapped_handle;
|
|
Status s;
|
|
{
|
|
std::lock_guard<std::mutex> lock(cache_res_mgr_mu_);
|
|
s = cache_res_mgr_->MakeCacheReservation(incremental_memory_used,
|
|
&wrapped_handle);
|
|
}
|
|
(*handle).reset(
|
|
new ConcurrentCacheReservationManager::CacheReservationHandle(
|
|
std::enable_shared_from_this<
|
|
ConcurrentCacheReservationManager>::shared_from_this(),
|
|
std::move(wrapped_handle)));
|
|
return s;
|
|
}
|
|
inline std::size_t GetTotalReservedCacheSize() override {
|
|
return cache_res_mgr_->GetTotalReservedCacheSize();
|
|
}
|
|
inline std::size_t GetTotalMemoryUsed() override {
|
|
std::lock_guard<std::mutex> lock(cache_res_mgr_mu_);
|
|
return cache_res_mgr_->GetTotalMemoryUsed();
|
|
}
|
|
|
|
private:
|
|
std::mutex cache_res_mgr_mu_;
|
|
std::shared_ptr<CacheReservationManager> cache_res_mgr_;
|
|
};
|
|
} // namespace ROCKSDB_NAMESPACE
|