mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-26 16:30:56 +00:00
204fcff751
Summary: Internally refactors SecondaryCache integration out of LRUCache specifically and into a wrapper/adapter class that works with various Cache implementations. Notably, this relies on separating the notion of async lookup handles from other cache handles, so that HyperClockCache doesn't have to deal with the problem of allocating handles from the hash table for lookups that might fail anyway, and might be on the same key without support for coalescing. (LRUCache's hash table can incorporate previously allocated handles thanks to its pointer indirection.) Specifically, I'm worried about the case in which hundreds of threads try to access the same block and probing in the hash table degrades to linear search on the pile of entries with the same key. This change is a big step in the direction of supporting stacked SecondaryCaches, but there are obstacles to completing that. Especially, there is no SecondaryCache hook for evictions to pass from one to the next. It has been proposed that evictions be transmitted simply as the persisted data (as in SaveToCallback), but given the current structure provided by the CacheItemHelpers, that would require an extra copy of the block data, because there's intentionally no way to ask for a contiguous Slice of the data (to allow for flexibility in storage). `AsyncLookupHandle` and the re-worked `WaitAll()` should be essentially prepared for stacked SecondaryCaches, but several "TODO with stacked secondaries" issues remain in various places. It could be argued that the stacking instead be done as a SecondaryCache adapter that wraps two (or more) SecondaryCaches, but at least with the current API that would require an extra heap allocation on SecondaryCache Lookup for a wrapper SecondaryCacheResultHandle that can transfer a Lookup between secondaries. We could also consider trying to unify the Cache and SecondaryCache APIs, though that might be difficult if `AsyncLookupHandle` is kept a fixed struct. ## cache.h (public API) Moves `secondary_cache` option from LRUCacheOptions to ShardedCacheOptions so that it is applicable to HyperClockCache. ## advanced_cache.h (advanced public API) * Add `Cache::CreateStandalone()` so that the SecondaryCache support wrapper can use it. * Add `SetEvictionCallback()` / `eviction_callback_` so that the SecondaryCache support wrapper can use it. Only a single callback is supported for efficiency. If there is ever a need for more than one, hopefully that can be handled with a broadcast callback wrapper. These are essentially the two "extra" pieces of `Cache` for pulling out specific SecondaryCache support from the `Cache` implementation. I think it's a good trade-off as these are reasonable, limited, and reusable "cut points" into the `Cache` implementations. * Remove async capability from standard `Lookup()` (getting rid of awkward restrictions on pending Handles) and add `AsyncLookupHandle` and `StartAsyncLookup()`. As noted in the comments, the full struct of `AsyncLookupHandle` is exposed so that it can be stack allocated, for efficiency, though more data is being copied around than before, which could impact performance. (Lookup info -> AsyncLookupHandle -> Handle vs. Lookup info -> Handle) I could foresee a future in which a Cache internally saves a pointer to the AsyncLookupHandle, which means it's dangerous to allow it to be copyable or even movable. It also means it's not compatible with std::vector (which I don't like requiring as an API parameter anyway), so `WaitAll()` expects any contiguous array of AsyncLookupHandles. I believe this is best for common case efficiency, while behaving well in other cases also. For example, `WaitAll()` has no effect on default-constructed AsyncLookupHandles, which look like a completed cache miss. ## cacheable_entry.h A couple of functions are obsolete because Cache::Handle can no longer be pending. ## cache.cc Provides default implementations for new or revamped Cache functions, especially appropriate for non-blocking caches. ## secondary_cache_adapter.{h,cc} The full details of the Cache wrapper adding SecondaryCache support. Essentially replicates the SecondaryCache handling that was in LRUCache, but obviously refactored. There is a bit of logic duplication, where Lookup() is essentially a manually optimized version of StartAsyncLookup() and Wait(), but it's roughly a dozen lines of code. ## sharded_cache.h, typed_cache.h, charged_cache.{h,cc}, sim_cache.cc Simply updated for Cache API changes. ## lru_cache.{h,cc} Carefully remove SecondaryCache logic, implement `CreateStandalone` and eviction handler functionality. ## clock_cache.{h,cc} Expose existing `CreateStandalone` functionality, add eviction handler functionality. Light refactoring. ## block_based_table_reader* Mostly re-worked the only usage of async Lookup, which is in BlockBasedTable::MultiGet. Used arrays in place of autovector in some places for efficiency. Simplified some logic by not trying to process some cache results before they're all ready. Created new function `BlockBasedTable::GetCachePriority()` to reduce some pre-existing code duplication (and avoid making it worse). Fixed at least one small bug from the prior confusing mixture of async and sync Lookups. In MaybeReadBlockAndLoadToCache(), called by RetrieveBlock(), called by MultiGet() with wait=false, is_cache_hit for the block_cache_tracer entry would not be set to true if the handle was pending after Lookup and before Wait. ## Intended follow-up work * Figure out if there are any missing stats or block_cache_tracer work in refactored BlockBasedTable::MultiGet * Stacked secondary caches (see above discussion) * See if we can make up for the small MultiGet performance regression. * Study more performance with SecondaryCache * Items evicted from over-full LRUCache in Release were not being demoted to SecondaryCache, and still aren't to minimize unit test churn. Ideally they would be demoted, but it's an exceptional case so not a big deal. * Use CreateStandalone for cache reservations (save unnecessary hash table operations). Not a big deal, but worthy cleanup. * Somehow I got the contract for SecondaryCache::Insert wrong in #10945. (Doesn't take ownership!) That API comment needs to be fixed, but didn't want to mingle that in here. Pull Request resolved: https://github.com/facebook/rocksdb/pull/11301 Test Plan: ## Unit tests Generally updated to include HCC in SecondaryCache tests, though HyperClockCache has some different, less strict behaviors that leads to some tests not really being set up to work with it. Some of the tests remain disabled with it, but I think we have good coverage without them. ## Crash/stress test Updated to use the new combination. ## Performance First, let's check for regression on caches without secondary cache configured. Adding support for the eviction callback is likely to have a tiny effect, but it shouldn't be worrisome. LRUCache could benefit slightly from less logic around SecondaryCache handling. We can test with cache_bench default settings, built with DEBUG_LEVEL=0 and PORTABLE=0. ``` (while :; do base/cache_bench --cache_type=hyper_clock_cache | grep Rough; done) | awk '{ sum += $9; count++; print $0; print "Average: " int(sum / count) }' ``` **Before** this and #11299 (which could also have a small effect), running for about an hour, before & after running concurrently for each cache type: HyperClockCache: 3168662 (average parallel ops/sec) LRUCache: 2940127 **After** this and #11299, running for about an hour: HyperClockCache: 3164862 (average parallel ops/sec) (0.12% slower) LRUCache: 2940928 (0.03% faster) This is an acceptable difference IMHO. Next, let's consider essentially the worst case of new CPU overhead affecting overall performance. MultiGet uses the async lookup interface regardless of whether SecondaryCache or folly are used. We can configure a benchmark where all block cache queries are for data blocks, and all are hits. Create DB and test (before and after tests running simultaneously): ``` TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=fillrandom -num=30000000 -disable_wal=1 -bloom_bits=16 TEST_TMPDIR=/dev/shm base/db_bench -benchmarks=multireadrandom[-X30] -readonly -multiread_batched -batch_size=32 -num=30000000 -bloom_bits=16 -cache_size=6789000000 -duration 20 -threads=16 ``` **Before**: multireadrandom [AVG 30 runs] : 3444202 (± 57049) ops/sec; 240.9 (± 4.0) MB/sec multireadrandom [MEDIAN 30 runs] : 3514443 ops/sec; 245.8 MB/sec **After**: multireadrandom [AVG 30 runs] : 3291022 (± 58851) ops/sec; 230.2 (± 4.1) MB/sec multireadrandom [MEDIAN 30 runs] : 3366179 ops/sec; 235.4 MB/sec So that's roughly a 3% regression, on kind of a *worst case* test of MultiGet CPU. Similar story with HyperClockCache: **Before**: multireadrandom [AVG 30 runs] : 3933777 (± 41840) ops/sec; 275.1 (± 2.9) MB/sec multireadrandom [MEDIAN 30 runs] : 3970667 ops/sec; 277.7 MB/sec **After**: multireadrandom [AVG 30 runs] : 3755338 (± 30391) ops/sec; 262.6 (± 2.1) MB/sec multireadrandom [MEDIAN 30 runs] : 3785696 ops/sec; 264.8 MB/sec Roughly a 4-5% regression. Not ideal, but not the whole story, fortunately. Let's also look at Get() in db_bench: ``` TEST_TMPDIR=/dev/shm ./db_bench -benchmarks=readrandom[-X30] -readonly -num=30000000 -bloom_bits=16 -cache_size=6789000000 -duration 20 -threads=16 ``` **Before**: readrandom [AVG 30 runs] : 2198685 (± 13412) ops/sec; 153.8 (± 0.9) MB/sec readrandom [MEDIAN 30 runs] : 2209498 ops/sec; 154.5 MB/sec **After**: readrandom [AVG 30 runs] : 2292814 (± 43508) ops/sec; 160.3 (± 3.0) MB/sec readrandom [MEDIAN 30 runs] : 2365181 ops/sec; 165.4 MB/sec That's showing roughly a 4% improvement, perhaps because of the secondary cache code that is no longer part of LRUCache. But weirdly, HyperClockCache is also showing 2-3% improvement: **Before**: readrandom [AVG 30 runs] : 2272333 (± 9992) ops/sec; 158.9 (± 0.7) MB/sec readrandom [MEDIAN 30 runs] : 2273239 ops/sec; 159.0 MB/sec **After**: readrandom [AVG 30 runs] : 2332407 (± 11252) ops/sec; 163.1 (± 0.8) MB/sec readrandom [MEDIAN 30 runs] : 2335329 ops/sec; 163.3 MB/sec Reviewed By: ltamasi Differential Revision: D44177044 Pulled By: pdillinger fbshipit-source-id: e808e48ff3fe2f792a79841ba617be98e48689f5
376 lines
14 KiB
C++
376 lines
14 KiB
C++
// Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
|
|
// APIs for accessing Cache in a type-safe and convenient way. Cache is kept
|
|
// at a low, thin level of abstraction so that different implementations can
|
|
// be plugged in, but these wrappers provide clean, convenient access to the
|
|
// most common operations.
|
|
//
|
|
// A number of template classes are needed for sharing common structure. The
|
|
// key classes are these:
|
|
//
|
|
// * PlaceholderCacheInterface - Used for making cache reservations, with
|
|
// entries that have a charge but no value.
|
|
// * BasicTypedCacheInterface<TValue> - Used for primary cache storage of
|
|
// objects of type TValue.
|
|
// * FullTypedCacheHelper<TValue, TCreateContext> - Used for secondary cache
|
|
// compatible storage of objects of type TValue.
|
|
// * For each of these, there's a "Shared" version
|
|
// (e.g. FullTypedSharedCacheInterface) that holds a shared_ptr to the Cache,
|
|
// rather than assuming external ownership by holding only a raw `Cache*`.
|
|
|
|
#pragma once
|
|
|
|
#include <algorithm>
|
|
#include <cstdint>
|
|
#include <memory>
|
|
#include <type_traits>
|
|
|
|
#include "cache/cache_helpers.h"
|
|
#include "rocksdb/advanced_cache.h"
|
|
#include "rocksdb/advanced_options.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
// For future consideration:
|
|
// * Pass in value to Insert with std::unique_ptr& to simplify ownership
|
|
// transfer logic in callers
|
|
// * Make key type a template parameter (e.g. useful for table cache)
|
|
// * Closer integration with CacheHandleGuard (opt-in, so not always
|
|
// paying the extra overhead)
|
|
|
|
#define CACHE_TYPE_DEFS() \
|
|
using Priority = Cache::Priority; \
|
|
using Handle = Cache::Handle; \
|
|
using ObjectPtr = Cache::ObjectPtr; \
|
|
using CreateContext = Cache::CreateContext; \
|
|
using CacheItemHelper = Cache::CacheItemHelper /* caller ; */
|
|
|
|
template <typename CachePtr>
|
|
class BaseCacheInterface {
|
|
public:
|
|
CACHE_TYPE_DEFS();
|
|
|
|
/*implicit*/ BaseCacheInterface(CachePtr cache) : cache_(std::move(cache)) {}
|
|
|
|
inline void Release(Handle* handle) { cache_->Release(handle); }
|
|
|
|
inline void ReleaseAndEraseIfLastRef(Handle* handle) {
|
|
cache_->Release(handle, /*erase_if_last_ref*/ true);
|
|
}
|
|
|
|
inline void RegisterReleaseAsCleanup(Handle* handle, Cleanable& cleanable) {
|
|
cleanable.RegisterCleanup(&ReleaseCacheHandleCleanup, get(), handle);
|
|
}
|
|
|
|
inline Cache* get() const { return &*cache_; }
|
|
|
|
explicit inline operator bool() const noexcept { return cache_ != nullptr; }
|
|
|
|
protected:
|
|
CachePtr cache_;
|
|
};
|
|
|
|
// PlaceholderCacheInterface - Used for making cache reservations, with
|
|
// entries that have a charge but no value. CacheEntryRole is required as
|
|
// a template parameter.
|
|
template <CacheEntryRole kRole, typename CachePtr = Cache*>
|
|
class PlaceholderCacheInterface : public BaseCacheInterface<CachePtr> {
|
|
public:
|
|
CACHE_TYPE_DEFS();
|
|
using BaseCacheInterface<CachePtr>::BaseCacheInterface;
|
|
|
|
inline Status Insert(const Slice& key, size_t charge, Handle** handle) {
|
|
return this->cache_->Insert(key, /*value=*/nullptr, GetHelper(), charge,
|
|
handle);
|
|
}
|
|
|
|
static const Cache::CacheItemHelper* GetHelper() {
|
|
static const Cache::CacheItemHelper kHelper{kRole};
|
|
return &kHelper;
|
|
}
|
|
};
|
|
|
|
template <CacheEntryRole kRole>
|
|
using PlaceholderSharedCacheInterface =
|
|
PlaceholderCacheInterface<kRole, std::shared_ptr<Cache>>;
|
|
|
|
template <class TValue>
|
|
class BasicTypedCacheHelperFns {
|
|
public:
|
|
CACHE_TYPE_DEFS();
|
|
// E.g. char* for char[]
|
|
using TValuePtr = std::remove_extent_t<TValue>*;
|
|
|
|
protected:
|
|
inline static ObjectPtr UpCastValue(TValuePtr value) { return value; }
|
|
inline static TValuePtr DownCastValue(ObjectPtr value) {
|
|
return static_cast<TValuePtr>(value);
|
|
}
|
|
|
|
static void Delete(ObjectPtr value, MemoryAllocator* allocator) {
|
|
// FIXME: Currently, no callers actually allocate the ObjectPtr objects
|
|
// using the custom allocator, just subobjects that keep a reference to
|
|
// the allocator themselves (with CacheAllocationPtr).
|
|
if (/*DISABLED*/ false && allocator) {
|
|
if constexpr (std::is_destructible_v<TValue>) {
|
|
DownCastValue(value)->~TValue();
|
|
}
|
|
allocator->Deallocate(value);
|
|
} else {
|
|
// Like delete but properly handles TValue=char[] etc.
|
|
std::default_delete<TValue>{}(DownCastValue(value));
|
|
}
|
|
}
|
|
};
|
|
|
|
// In its own class to try to minimize the number of distinct CacheItemHelper
|
|
// instances (e.g. don't vary by CachePtr)
|
|
template <class TValue, CacheEntryRole kRole>
|
|
class BasicTypedCacheHelper : public BasicTypedCacheHelperFns<TValue> {
|
|
public:
|
|
static const Cache::CacheItemHelper* GetBasicHelper() {
|
|
static const Cache::CacheItemHelper kHelper{kRole,
|
|
&BasicTypedCacheHelper::Delete};
|
|
return &kHelper;
|
|
}
|
|
};
|
|
|
|
// BasicTypedCacheInterface - Used for primary cache storage of objects of
|
|
// type TValue, which can be cleaned up with std::default_delete<TValue>. The
|
|
// role is provided by TValue::kCacheEntryRole or given in an optional
|
|
// template parameter.
|
|
template <class TValue, CacheEntryRole kRole = TValue::kCacheEntryRole,
|
|
typename CachePtr = Cache*>
|
|
class BasicTypedCacheInterface : public BaseCacheInterface<CachePtr>,
|
|
public BasicTypedCacheHelper<TValue, kRole> {
|
|
public:
|
|
CACHE_TYPE_DEFS();
|
|
using typename BasicTypedCacheHelperFns<TValue>::TValuePtr;
|
|
struct TypedHandle : public Handle {};
|
|
using BasicTypedCacheHelper<TValue, kRole>::GetBasicHelper;
|
|
// ctor
|
|
using BaseCacheInterface<CachePtr>::BaseCacheInterface;
|
|
struct TypedAsyncLookupHandle : public Cache::AsyncLookupHandle {
|
|
TypedHandle* Result() {
|
|
return reinterpret_cast<TypedHandle*>(Cache::AsyncLookupHandle::Result());
|
|
}
|
|
};
|
|
|
|
inline Status Insert(const Slice& key, TValuePtr value, size_t charge,
|
|
TypedHandle** handle = nullptr,
|
|
Priority priority = Priority::LOW) {
|
|
auto untyped_handle = reinterpret_cast<Handle**>(handle);
|
|
return this->cache_->Insert(
|
|
key, BasicTypedCacheHelperFns<TValue>::UpCastValue(value),
|
|
GetBasicHelper(), charge, untyped_handle, priority);
|
|
}
|
|
|
|
inline TypedHandle* Lookup(const Slice& key, Statistics* stats = nullptr) {
|
|
return reinterpret_cast<TypedHandle*>(
|
|
this->cache_->BasicLookup(key, stats));
|
|
}
|
|
|
|
inline void StartAsyncLookup(TypedAsyncLookupHandle& async_handle) {
|
|
assert(async_handle.helper == nullptr);
|
|
this->cache_->StartAsyncLookup(async_handle);
|
|
}
|
|
|
|
inline CacheHandleGuard<TValue> Guard(TypedHandle* handle) {
|
|
if (handle) {
|
|
return CacheHandleGuard<TValue>(&*this->cache_, handle);
|
|
} else {
|
|
return {};
|
|
}
|
|
}
|
|
|
|
inline std::shared_ptr<TValue> SharedGuard(TypedHandle* handle) {
|
|
if (handle) {
|
|
return MakeSharedCacheHandleGuard<TValue>(&*this->cache_, handle);
|
|
} else {
|
|
return {};
|
|
}
|
|
}
|
|
|
|
inline TValuePtr Value(TypedHandle* handle) {
|
|
return BasicTypedCacheHelperFns<TValue>::DownCastValue(
|
|
this->cache_->Value(handle));
|
|
}
|
|
};
|
|
|
|
// BasicTypedSharedCacheInterface - Like BasicTypedCacheInterface but with a
|
|
// shared_ptr<Cache> for keeping Cache alive.
|
|
template <class TValue, CacheEntryRole kRole = TValue::kCacheEntryRole>
|
|
using BasicTypedSharedCacheInterface =
|
|
BasicTypedCacheInterface<TValue, kRole, std::shared_ptr<Cache>>;
|
|
|
|
// TValue must implement ContentSlice() and ~TValue
|
|
// TCreateContext must implement Create(std::unique_ptr<TValue>*, ...)
|
|
template <class TValue, class TCreateContext>
|
|
class FullTypedCacheHelperFns : public BasicTypedCacheHelperFns<TValue> {
|
|
public:
|
|
CACHE_TYPE_DEFS();
|
|
|
|
protected:
|
|
using typename BasicTypedCacheHelperFns<TValue>::TValuePtr;
|
|
using BasicTypedCacheHelperFns<TValue>::DownCastValue;
|
|
using BasicTypedCacheHelperFns<TValue>::UpCastValue;
|
|
|
|
static size_t Size(ObjectPtr v) {
|
|
TValuePtr value = DownCastValue(v);
|
|
auto slice = value->ContentSlice();
|
|
return slice.size();
|
|
}
|
|
|
|
static Status SaveTo(ObjectPtr v, size_t from_offset, size_t length,
|
|
char* out) {
|
|
TValuePtr value = DownCastValue(v);
|
|
auto slice = value->ContentSlice();
|
|
assert(from_offset < slice.size());
|
|
assert(from_offset + length <= slice.size());
|
|
std::copy_n(slice.data() + from_offset, length, out);
|
|
return Status::OK();
|
|
}
|
|
|
|
static Status Create(const Slice& data, CreateContext* context,
|
|
MemoryAllocator* allocator, ObjectPtr* out_obj,
|
|
size_t* out_charge) {
|
|
std::unique_ptr<TValue> value = nullptr;
|
|
if constexpr (sizeof(TCreateContext) > 0) {
|
|
TCreateContext* tcontext = static_cast<TCreateContext*>(context);
|
|
tcontext->Create(&value, out_charge, data, allocator);
|
|
} else {
|
|
TCreateContext::Create(&value, out_charge, data, allocator);
|
|
}
|
|
*out_obj = UpCastValue(value.release());
|
|
return Status::OK();
|
|
}
|
|
};
|
|
|
|
// In its own class to try to minimize the number of distinct CacheItemHelper
|
|
// instances (e.g. don't vary by CachePtr)
|
|
template <class TValue, class TCreateContext, CacheEntryRole kRole>
|
|
class FullTypedCacheHelper
|
|
: public FullTypedCacheHelperFns<TValue, TCreateContext> {
|
|
public:
|
|
static const Cache::CacheItemHelper* GetFullHelper() {
|
|
static const Cache::CacheItemHelper kHelper{
|
|
kRole,
|
|
&FullTypedCacheHelper::Delete,
|
|
&FullTypedCacheHelper::Size,
|
|
&FullTypedCacheHelper::SaveTo,
|
|
&FullTypedCacheHelper::Create,
|
|
BasicTypedCacheHelper<TValue, kRole>::GetBasicHelper()};
|
|
return &kHelper;
|
|
}
|
|
};
|
|
|
|
// FullTypedCacheHelper - Used for secondary cache compatible storage of
|
|
// objects of type TValue. In addition to BasicTypedCacheInterface constraints,
|
|
// we require TValue::ContentSlice() to return persistable data. This
|
|
// simplifies usage for the normal case of simple secondary cache compatibility
|
|
// (can give you a Slice to the data already in memory). In addition to
|
|
// TCreateContext performing the role of Cache::CreateContext, it is also
|
|
// expected to provide a function Create(std::unique_ptr<TValue>* value,
|
|
// size_t* out_charge, const Slice& data, MemoryAllocator* allocator) for
|
|
// creating new TValue.
|
|
template <class TValue, class TCreateContext,
|
|
CacheEntryRole kRole = TValue::kCacheEntryRole,
|
|
typename CachePtr = Cache*>
|
|
class FullTypedCacheInterface
|
|
: public BasicTypedCacheInterface<TValue, kRole, CachePtr>,
|
|
public FullTypedCacheHelper<TValue, TCreateContext, kRole> {
|
|
public:
|
|
CACHE_TYPE_DEFS();
|
|
using typename BasicTypedCacheInterface<TValue, kRole, CachePtr>::TypedHandle;
|
|
using typename BasicTypedCacheInterface<TValue, kRole,
|
|
CachePtr>::TypedAsyncLookupHandle;
|
|
using typename BasicTypedCacheHelperFns<TValue>::TValuePtr;
|
|
using BasicTypedCacheHelper<TValue, kRole>::GetBasicHelper;
|
|
using FullTypedCacheHelper<TValue, TCreateContext, kRole>::GetFullHelper;
|
|
using BasicTypedCacheHelperFns<TValue>::UpCastValue;
|
|
using BasicTypedCacheHelperFns<TValue>::DownCastValue;
|
|
// ctor
|
|
using BasicTypedCacheInterface<TValue, kRole,
|
|
CachePtr>::BasicTypedCacheInterface;
|
|
|
|
// Insert with SecondaryCache compatibility (subject to CacheTier).
|
|
// (Basic Insert() also inherited.)
|
|
inline Status InsertFull(
|
|
const Slice& key, TValuePtr value, size_t charge,
|
|
TypedHandle** handle = nullptr, Priority priority = Priority::LOW,
|
|
CacheTier lowest_used_cache_tier = CacheTier::kNonVolatileBlockTier) {
|
|
auto untyped_handle = reinterpret_cast<Handle**>(handle);
|
|
auto helper = lowest_used_cache_tier == CacheTier::kNonVolatileBlockTier
|
|
? GetFullHelper()
|
|
: GetBasicHelper();
|
|
return this->cache_->Insert(key, UpCastValue(value), helper, charge,
|
|
untyped_handle, priority);
|
|
}
|
|
|
|
// Like SecondaryCache::InsertSaved, with SecondaryCache compatibility
|
|
// (subject to CacheTier).
|
|
inline Status InsertSaved(
|
|
const Slice& key, const Slice& data, TCreateContext* create_context,
|
|
Priority priority = Priority::LOW,
|
|
CacheTier lowest_used_cache_tier = CacheTier::kNonVolatileBlockTier,
|
|
size_t* out_charge = nullptr) {
|
|
ObjectPtr value;
|
|
size_t charge;
|
|
Status st = GetFullHelper()->create_cb(data, create_context,
|
|
this->cache_->memory_allocator(),
|
|
&value, &charge);
|
|
if (out_charge) {
|
|
*out_charge = charge;
|
|
}
|
|
if (st.ok()) {
|
|
st = InsertFull(key, DownCastValue(value), charge, nullptr /*handle*/,
|
|
priority, lowest_used_cache_tier);
|
|
} else {
|
|
GetFullHelper()->del_cb(value, this->cache_->memory_allocator());
|
|
}
|
|
return st;
|
|
}
|
|
|
|
// Lookup with SecondaryCache support (subject to CacheTier).
|
|
// (Basic Lookup() also inherited.)
|
|
inline TypedHandle* LookupFull(
|
|
const Slice& key, TCreateContext* create_context = nullptr,
|
|
Priority priority = Priority::LOW, Statistics* stats = nullptr,
|
|
CacheTier lowest_used_cache_tier = CacheTier::kNonVolatileBlockTier) {
|
|
if (lowest_used_cache_tier == CacheTier::kNonVolatileBlockTier) {
|
|
return reinterpret_cast<TypedHandle*>(this->cache_->Lookup(
|
|
key, GetFullHelper(), create_context, priority, stats));
|
|
} else {
|
|
return BasicTypedCacheInterface<TValue, kRole, CachePtr>::Lookup(key,
|
|
stats);
|
|
}
|
|
}
|
|
|
|
inline void StartAsyncLookupFull(
|
|
TypedAsyncLookupHandle& async_handle,
|
|
CacheTier lowest_used_cache_tier = CacheTier::kNonVolatileBlockTier) {
|
|
if (lowest_used_cache_tier == CacheTier::kNonVolatileBlockTier) {
|
|
async_handle.helper = GetFullHelper();
|
|
this->cache_->StartAsyncLookup(async_handle);
|
|
} else {
|
|
BasicTypedCacheInterface<TValue, kRole, CachePtr>::StartAsyncLookup(
|
|
async_handle);
|
|
}
|
|
}
|
|
};
|
|
|
|
// FullTypedSharedCacheInterface - Like FullTypedCacheInterface but with a
|
|
// shared_ptr<Cache> for keeping Cache alive.
|
|
template <class TValue, class TCreateContext,
|
|
CacheEntryRole kRole = TValue::kCacheEntryRole>
|
|
using FullTypedSharedCacheInterface =
|
|
FullTypedCacheInterface<TValue, TCreateContext, kRole,
|
|
std::shared_ptr<Cache>>;
|
|
|
|
#undef CACHE_TYPE_DEFS
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|