mirror of https://github.com/facebook/rocksdb.git
Put Cache and CacheWrapper in new public header (#11192)
Summary: The definition of the Cache class should not be needed by the vast majority of RocksDB users, so I think it is just distracting to include it in cache.h, which is primarily needed for configuring and creating caches. This change moves the class to a new header advanced_cache.h. It is just cut-and-paste except for modifying the class API comment. In general, operations on shared_ptr<Cache> should continue to work when only a forward declaration of Cache is available, as long as all the Cache instances provided are already shared_ptr. See https://stackoverflow.com/a/17650101/454544 Also, the most common way to customize a Cache is by wrapping an existing implementation, so it makes sense to provide CacheWrapper in the public API. This was a cut-and-paste job except removing the implementation of Name() so that derived classes must provide it. Intended follow-up: consolidate Release() into one function to reduce customization bugs / confusion Pull Request resolved: https://github.com/facebook/rocksdb/pull/11192 Test Plan: `make check` Reviewed By: anand1976 Differential Revision: D43055487 Pulled By: pdillinger fbshipit-source-id: 7b05492df35e0f30b581b4c24c579bc275b6d110
This commit is contained in:
parent
b7747bbc9f
commit
3cacd4b4ec
|
@ -22,6 +22,7 @@
|
|||
* Removed the deprecated version of these utility functions and the corresponding Java bindings: `LoadOptionsFromFile`, `LoadLatestOptions`, `CheckOptionsCompatibility`.
|
||||
|
||||
### Public API Changes
|
||||
* Moved rarely-needed Cache class definition to new advanced_cache.h, and added a CacheWrapper class to advanced_cache.h. Minor changes to SimCache API definitions.
|
||||
* Completely removed the following deprecated/obsolete statistics: the tickers `BLOCK_CACHE_INDEX_BYTES_EVICT`, `BLOCK_CACHE_FILTER_BYTES_EVICT`, `BLOOM_FILTER_MICROS`, `NO_FILE_CLOSES`, `STALL_L0_SLOWDOWN_MICROS`, `STALL_MEMTABLE_COMPACTION_MICROS`, `STALL_L0_NUM_FILES_MICROS`, `RATE_LIMIT_DELAY_MILLIS`, `NO_ITERATORS`, `NUMBER_FILTERED_DELETES`, `WRITE_TIMEDOUT`, `BLOB_DB_GC_NUM_KEYS_OVERWRITTEN`, `BLOB_DB_GC_NUM_KEYS_EXPIRED`, `BLOB_DB_GC_BYTES_OVERWRITTEN`, `BLOB_DB_GC_BYTES_EXPIRED`, `BLOCK_CACHE_COMPRESSION_DICT_BYTES_EVICT` as well as the histograms `STALL_L0_SLOWDOWN_COUNT`, `STALL_MEMTABLE_COMPACTION_COUNT`, `STALL_L0_NUM_FILES_COUNT`, `HARD_RATE_LIMIT_DELAY_COUNT`, `SOFT_RATE_LIMIT_DELAY_COUNT`, `BLOB_DB_GC_MICROS`, and `NUM_DATA_BLOCKS_READ_PER_LEVEL`. Note that as a result, the C++ enum values of the still supported statistics have changed. Developers are advised to not rely on the actual numeric values.
|
||||
* Deprecated IngestExternalFileOptions::write_global_seqno and change default to false. This option only needs to be set to true to generate a DB compatible with RocksDB versions before 5.16.0.
|
||||
* Remove deprecated APIs `GetColumnFamilyOptionsFrom{Map|String}(const ColumnFamilyOptions&, ..)`, `GetDBOptionsFrom{Map|String}(const DBOptions&, ..)`, `GetBlockBasedTableOptionsFrom{Map|String}(const BlockBasedTableOptions& table_options, ..)` and ` GetPlainTableOptionsFrom{Map|String}(const PlainTableOptions& table_options,..)`.
|
||||
|
|
|
@ -16,7 +16,7 @@
|
|||
#include "db/db_impl/db_impl.h"
|
||||
#include "monitoring/histogram.h"
|
||||
#include "port/port.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/convenience.h"
|
||||
#include "rocksdb/db.h"
|
||||
#include "rocksdb/env.h"
|
||||
|
|
|
@ -7,7 +7,7 @@
|
|||
|
||||
#include <cassert>
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/rocksdb_namespace.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
|
|
@ -8,7 +8,7 @@
|
|||
#include <algorithm>
|
||||
#include <atomic>
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "table/unique_id_impl.h"
|
||||
#include "util/hash.h"
|
||||
#include "util/math.h"
|
||||
|
|
|
@ -8,7 +8,7 @@
|
|||
#include <string>
|
||||
|
||||
#include "port/port.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
||||
|
|
|
@ -15,7 +15,7 @@
|
|||
|
||||
#include "port/lang.h"
|
||||
#include "port/port.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "util/hash.h"
|
||||
#include "util/mutexlock.h"
|
||||
|
||||
|
|
|
@ -29,8 +29,8 @@
|
|||
#include <type_traits>
|
||||
|
||||
#include "cache/cache_helpers.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/advanced_options.h"
|
||||
#include "rocksdb/cache.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
||||
|
|
|
@ -8,7 +8,7 @@
|
|||
#include <memory>
|
||||
|
||||
#include "memory/memory_allocator.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/rocksdb_namespace.h"
|
||||
#include "rocksdb/slice.h"
|
||||
#include "rocksdb/status.h"
|
||||
|
|
3
db/c.cc
3
db/c.cc
|
@ -16,7 +16,7 @@
|
|||
#include <vector>
|
||||
|
||||
#include "port/port.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/compaction_filter.h"
|
||||
#include "rocksdb/comparator.h"
|
||||
#include "rocksdb/convenience.h"
|
||||
|
@ -6397,4 +6397,3 @@ void rocksdb_enable_manual_compaction(rocksdb_t* db) {
|
|||
}
|
||||
|
||||
} // end extern "C"
|
||||
|
||||
|
|
|
@ -287,9 +287,11 @@ class PersistentCacheFromCache : public PersistentCache {
|
|||
};
|
||||
|
||||
class ReadOnlyCacheWrapper : public CacheWrapper {
|
||||
public:
|
||||
using CacheWrapper::CacheWrapper;
|
||||
|
||||
using Cache::Insert;
|
||||
const char* Name() const override { return "ReadOnlyCacheWrapper"; }
|
||||
|
||||
Status Insert(const Slice& /*key*/, Cache::ObjectPtr /*value*/,
|
||||
const CacheItemHelper* /*helper*/, size_t /*charge*/,
|
||||
Handle** /*handle*/, Priority /*priority*/) override {
|
||||
|
@ -711,7 +713,8 @@ class LookupLiarCache : public CacheWrapper {
|
|||
explicit LookupLiarCache(std::shared_ptr<Cache> target)
|
||||
: CacheWrapper(std::move(target)) {}
|
||||
|
||||
using Cache::Lookup;
|
||||
const char* Name() const override { return "LookupLiarCache"; }
|
||||
|
||||
Handle* Lookup(const Slice& key, const CacheItemHelper* helper = nullptr,
|
||||
CreateContext* create_context = nullptr,
|
||||
Priority priority = Priority::LOW, bool wait = true,
|
||||
|
|
|
@ -909,82 +909,6 @@ class TestPutOperator : public MergeOperator {
|
|||
virtual const char* Name() const override { return "TestPutOperator"; }
|
||||
};
|
||||
|
||||
// A wrapper around Cache that can easily be extended with instrumentation,
|
||||
// etc.
|
||||
class CacheWrapper : public Cache {
|
||||
public:
|
||||
explicit CacheWrapper(std::shared_ptr<Cache> target)
|
||||
: target_(std::move(target)) {}
|
||||
|
||||
const char* Name() const override { return target_->Name(); }
|
||||
|
||||
Status Insert(const Slice& key, ObjectPtr value,
|
||||
const CacheItemHelper* helper, size_t charge,
|
||||
Handle** handle = nullptr,
|
||||
Priority priority = Priority::LOW) override {
|
||||
return target_->Insert(key, value, helper, charge, handle, priority);
|
||||
}
|
||||
|
||||
Handle* Lookup(const Slice& key, const CacheItemHelper* helper,
|
||||
CreateContext* create_context,
|
||||
Priority priority = Priority::LOW, bool wait = true,
|
||||
Statistics* stats = nullptr) override {
|
||||
return target_->Lookup(key, helper, create_context, priority, wait, stats);
|
||||
}
|
||||
|
||||
bool Ref(Handle* handle) override { return target_->Ref(handle); }
|
||||
|
||||
using Cache::Release;
|
||||
bool Release(Handle* handle, bool erase_if_last_ref = false) override {
|
||||
return target_->Release(handle, erase_if_last_ref);
|
||||
}
|
||||
|
||||
ObjectPtr Value(Handle* handle) override { return target_->Value(handle); }
|
||||
|
||||
void Erase(const Slice& key) override { target_->Erase(key); }
|
||||
uint64_t NewId() override { return target_->NewId(); }
|
||||
|
||||
void SetCapacity(size_t capacity) override { target_->SetCapacity(capacity); }
|
||||
|
||||
void SetStrictCapacityLimit(bool strict_capacity_limit) override {
|
||||
target_->SetStrictCapacityLimit(strict_capacity_limit);
|
||||
}
|
||||
|
||||
bool HasStrictCapacityLimit() const override {
|
||||
return target_->HasStrictCapacityLimit();
|
||||
}
|
||||
|
||||
size_t GetCapacity() const override { return target_->GetCapacity(); }
|
||||
|
||||
size_t GetUsage() const override { return target_->GetUsage(); }
|
||||
|
||||
size_t GetUsage(Handle* handle) const override {
|
||||
return target_->GetUsage(handle);
|
||||
}
|
||||
|
||||
size_t GetPinnedUsage() const override { return target_->GetPinnedUsage(); }
|
||||
|
||||
size_t GetCharge(Handle* handle) const override {
|
||||
return target_->GetCharge(handle);
|
||||
}
|
||||
|
||||
const CacheItemHelper* GetCacheItemHelper(Handle* handle) const override {
|
||||
return target_->GetCacheItemHelper(handle);
|
||||
}
|
||||
|
||||
void ApplyToAllEntries(
|
||||
const std::function<void(const Slice& key, ObjectPtr value, size_t charge,
|
||||
const CacheItemHelper* helper)>& callback,
|
||||
const ApplyToAllEntriesOptions& opts) override {
|
||||
target_->ApplyToAllEntries(callback, opts);
|
||||
}
|
||||
|
||||
void EraseUnRefEntries() override { target_->EraseUnRefEntries(); }
|
||||
|
||||
protected:
|
||||
std::shared_ptr<Cache> target_;
|
||||
};
|
||||
|
||||
/*
|
||||
* A cache wrapper that tracks certain CacheEntryRole's cache charge, its
|
||||
* peaks and increments
|
||||
|
@ -1002,6 +926,8 @@ class TargetCacheChargeTrackingCache : public CacheWrapper {
|
|||
public:
|
||||
explicit TargetCacheChargeTrackingCache(std::shared_ptr<Cache> target);
|
||||
|
||||
const char* Name() const override { return "TargetCacheChargeTrackingCache"; }
|
||||
|
||||
Status Insert(const Slice& key, ObjectPtr value,
|
||||
const CacheItemHelper* helper, size_t charge,
|
||||
Handle** handle = nullptr,
|
||||
|
|
|
@ -20,8 +20,8 @@
|
|||
#include "db/wal_edit.h"
|
||||
#include "memory/arena.h"
|
||||
#include "port/malloc.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/advanced_options.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "table/table_reader.h"
|
||||
#include "table/unique_id_impl.h"
|
||||
#include "util/autovector.h"
|
||||
|
|
|
@ -0,0 +1,527 @@
|
|||
// Copyright (c) Meta Platforms, Inc. and affiliates.
|
||||
// This source code is licensed under both the GPLv2 (found in the
|
||||
// COPYING file in the root directory) and Apache 2.0 License
|
||||
// (found in the LICENSE.Apache file in the root directory).
|
||||
//
|
||||
// APIs for customizing read caches in RocksDB.
|
||||
|
||||
#pragma once
|
||||
|
||||
#include <cstdint>
|
||||
#include <functional>
|
||||
#include <memory>
|
||||
#include <string>
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/memory_allocator.h"
|
||||
#include "rocksdb/slice.h"
|
||||
#include "rocksdb/status.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
||||
class Logger;
|
||||
class Statistics;
|
||||
|
||||
// A Cache maps keys to objects resident in memory, tracks reference counts
|
||||
// on those key-object entries, and is able to remove unreferenced entries
|
||||
// whenever it wants. All operations are fully thread safe except as noted.
|
||||
// Inserted entries have a specified "charge" which is some quantity in
|
||||
// unspecified units, typically bytes of memory used. A Cache will typically
|
||||
// have a finite capacity in units of charge, and evict entries as needed
|
||||
// to stay at or below that capacity.
|
||||
//
|
||||
// NOTE: This API is for expert use only and is intended more for customizing
|
||||
// cache behavior than for calling into outside of RocksDB. It is subject to
|
||||
// change as RocksDB evolves, especially the RocksDB block cache. Overriding
|
||||
// CacheWrapper is the preferred way of customizing some operations on an
|
||||
// existing implementation.
|
||||
//
|
||||
// INTERNAL: See typed_cache.h for convenient wrappers on top of this API.
|
||||
// New virtual functions must also be added to CacheWrapper below.
|
||||
class Cache {
|
||||
public: // types hidden from API client
|
||||
// Opaque handle to an entry stored in the cache.
|
||||
struct Handle {};
|
||||
|
||||
public: // types hidden from Cache implementation
|
||||
// Pointer to cached object of unspecified type. (This type alias is
|
||||
// provided for clarity, not really for type checking.)
|
||||
using ObjectPtr = void*;
|
||||
|
||||
// Opaque object providing context (settings, etc.) to create objects
|
||||
// for primary cache from saved (serialized) secondary cache entries.
|
||||
struct CreateContext {};
|
||||
|
||||
public: // type defs
|
||||
// Depending on implementation, cache entries with higher priority levels
|
||||
// could be less likely to get evicted than entries with lower priority
|
||||
// levels. The "high" priority level applies to certain SST metablocks (e.g.
|
||||
// index and filter blocks) if the option
|
||||
// cache_index_and_filter_blocks_with_high_priority is set. The "low" priority
|
||||
// level is used for other kinds of SST blocks (most importantly, data
|
||||
// blocks), as well as the above metablocks in case
|
||||
// cache_index_and_filter_blocks_with_high_priority is
|
||||
// not set. The "bottom" priority level is for BlobDB's blob values.
|
||||
enum class Priority { HIGH, LOW, BOTTOM };
|
||||
|
||||
// A set of callbacks to allow objects in the primary block cache to be
|
||||
// be persisted in a secondary cache. The purpose of the secondary cache
|
||||
// is to support other ways of caching the object, such as persistent or
|
||||
// compressed data, that may require the object to be parsed and transformed
|
||||
// in some way. Since the primary cache holds C++ objects and the secondary
|
||||
// cache may only hold flat data that doesn't need relocation, these
|
||||
// callbacks need to be provided by the user of the block
|
||||
// cache to do the conversion.
|
||||
// The CacheItemHelper is passed to Insert() and Lookup(). It has pointers
|
||||
// to callback functions for size, saving and deletion of the
|
||||
// object. The callbacks are defined in C-style in order to make them
|
||||
// stateless and not add to the cache metadata size.
|
||||
// Saving multiple std::function objects will take up 32 bytes per
|
||||
// function, even if its not bound to an object and does no capture.
|
||||
//
|
||||
// All the callbacks are C-style function pointers in order to simplify
|
||||
// lifecycle management. Objects in the cache can outlive the parent DB,
|
||||
// so anything required for these operations should be contained in the
|
||||
// object itself.
|
||||
//
|
||||
// The SizeCallback takes a pointer to the object and returns the size
|
||||
// of the persistable data. It can be used by the secondary cache to allocate
|
||||
// memory if needed.
|
||||
//
|
||||
// RocksDB callbacks are NOT exception-safe. A callback completing with an
|
||||
// exception can lead to undefined behavior in RocksDB, including data loss,
|
||||
// unreported corruption, deadlocks, and more.
|
||||
using SizeCallback = size_t (*)(ObjectPtr obj);
|
||||
|
||||
// The SaveToCallback takes an object pointer and saves the persistable
|
||||
// data into a buffer. The secondary cache may decide to not store it in a
|
||||
// contiguous buffer, in which case this callback will be called multiple
|
||||
// times with increasing offset
|
||||
using SaveToCallback = Status (*)(ObjectPtr from_obj, size_t from_offset,
|
||||
size_t length, char* out_buf);
|
||||
|
||||
// A function pointer type for destruction of a cache object. This will
|
||||
// typically call the destructor for the appropriate type of the object.
|
||||
// The Cache is responsible for copying and reclaiming space for the key,
|
||||
// but objects are managed in part using this callback. Generally a DeleterFn
|
||||
// can be nullptr if the ObjectPtr does not need destruction (e.g. nullptr or
|
||||
// pointer into static data).
|
||||
using DeleterFn = void (*)(ObjectPtr obj, MemoryAllocator* allocator);
|
||||
|
||||
// The CreateCallback is takes in a buffer from the NVM cache and constructs
|
||||
// an object using it. The callback doesn't have ownership of the buffer and
|
||||
// should copy the contents into its own buffer. The CreateContext* is
|
||||
// provided by Lookup and may be used to follow DB- or CF-specific settings.
|
||||
// In case of some error, non-OK is returned and the caller should ignore
|
||||
// any result in out_obj. (The implementation must clean up after itself.)
|
||||
using CreateCallback = Status (*)(const Slice& data, CreateContext* context,
|
||||
MemoryAllocator* allocator,
|
||||
ObjectPtr* out_obj, size_t* out_charge);
|
||||
|
||||
// A struct with pointers to helper functions for spilling items from the
|
||||
// cache into the secondary cache. May be extended in the future. An
|
||||
// instance of this struct is expected to outlive the cache.
|
||||
struct CacheItemHelper {
|
||||
// Function for deleting an object on its removal from the Cache.
|
||||
// nullptr is only for entries that require no destruction, such as
|
||||
// "placeholder" cache entries with nullptr object.
|
||||
DeleterFn del_cb; // (<- Most performance critical)
|
||||
// Next three are used for persisting values as described above.
|
||||
// If any is nullptr, then all three should be nullptr and persisting the
|
||||
// entry to/from secondary cache is not supported.
|
||||
SizeCallback size_cb;
|
||||
SaveToCallback saveto_cb;
|
||||
CreateCallback create_cb;
|
||||
// Classification of the entry for monitoring purposes in block cache.
|
||||
CacheEntryRole role;
|
||||
|
||||
constexpr CacheItemHelper()
|
||||
: del_cb(nullptr),
|
||||
size_cb(nullptr),
|
||||
saveto_cb(nullptr),
|
||||
create_cb(nullptr),
|
||||
role(CacheEntryRole::kMisc) {}
|
||||
|
||||
explicit constexpr CacheItemHelper(CacheEntryRole _role,
|
||||
DeleterFn _del_cb = nullptr,
|
||||
SizeCallback _size_cb = nullptr,
|
||||
SaveToCallback _saveto_cb = nullptr,
|
||||
CreateCallback _create_cb = nullptr)
|
||||
: del_cb(_del_cb),
|
||||
size_cb(_size_cb),
|
||||
saveto_cb(_saveto_cb),
|
||||
create_cb(_create_cb),
|
||||
role(_role) {
|
||||
// Either all three secondary cache callbacks are non-nullptr or
|
||||
// all three are nullptr
|
||||
assert((size_cb != nullptr) == (saveto_cb != nullptr));
|
||||
assert((size_cb != nullptr) == (create_cb != nullptr));
|
||||
}
|
||||
inline bool IsSecondaryCacheCompatible() const {
|
||||
return size_cb != nullptr;
|
||||
}
|
||||
};
|
||||
|
||||
public: // ctor/dtor/create
|
||||
Cache(std::shared_ptr<MemoryAllocator> allocator = nullptr)
|
||||
: memory_allocator_(std::move(allocator)) {}
|
||||
// No copying allowed
|
||||
Cache(const Cache&) = delete;
|
||||
Cache& operator=(const Cache&) = delete;
|
||||
|
||||
// Destroys all remaining entries by calling the associated "deleter"
|
||||
virtual ~Cache() {}
|
||||
|
||||
// Creates a new Cache based on the input value string and returns the result.
|
||||
// Currently, this method can be used to create LRUCaches only
|
||||
// @param config_options
|
||||
// @param value The value might be:
|
||||
// - an old-style cache ("1M") -- equivalent to NewLRUCache(1024*102(
|
||||
// - Name-value option pairs -- "capacity=1M; num_shard_bits=4;
|
||||
// For the LRUCache, the values are defined in LRUCacheOptions.
|
||||
// @param result The new Cache object
|
||||
// @return OK if the cache was successfully created
|
||||
// @return NotFound if an invalid name was specified in the value
|
||||
// @return InvalidArgument if either the options were not valid
|
||||
static Status CreateFromString(const ConfigOptions& config_options,
|
||||
const std::string& value,
|
||||
std::shared_ptr<Cache>* result);
|
||||
|
||||
public: // functions
|
||||
// The type of the Cache
|
||||
virtual const char* Name() const = 0;
|
||||
|
||||
// The Insert and Lookup APIs below are intended to allow cached objects
|
||||
// to be demoted/promoted between the primary block cache and a secondary
|
||||
// cache. The secondary cache could be a non-volatile cache, and will
|
||||
// likely store the object in a different representation. They rely on a
|
||||
// per object CacheItemHelper to do the conversions.
|
||||
// The secondary cache may persist across process and system restarts,
|
||||
// and may even be moved between hosts. Therefore, the cache key must
|
||||
// be repeatable across restarts/reboots, and globally unique if
|
||||
// multiple DBs share the same cache and the set of DBs can change
|
||||
// over time.
|
||||
|
||||
// Insert a mapping from key->object into the cache and assign it
|
||||
// the specified charge against the total cache capacity. If
|
||||
// strict_capacity_limit is true and cache reaches its full capacity,
|
||||
// return Status::MemoryLimit. `obj` must be non-nullptr if compatible
|
||||
// with secondary cache (helper->size_cb != nullptr), because Value() ==
|
||||
// nullptr is reserved for indicating some secondary cache failure cases.
|
||||
// On success, returns OK and takes ownership of `obj`, eventually deleting
|
||||
// it with helper->del_cb. On non-OK return, the caller maintains ownership
|
||||
// of `obj` so will often need to delete it in such cases.
|
||||
//
|
||||
// The helper argument is saved by the cache and will be used when the
|
||||
// inserted object is evicted or considered for promotion to the secondary
|
||||
// cache. Promotion to secondary cache is only enabled if helper->size_cb
|
||||
// != nullptr. The helper must outlive the cache. Callers may use
|
||||
// &kNoopCacheItemHelper as a trivial helper (no deleter for the object,
|
||||
// no secondary cache). `helper` must not be nullptr (efficiency).
|
||||
//
|
||||
// If `handle` is not nullptr and return status is OK, `handle` is set
|
||||
// to a Handle* for the entry. The caller must call this->Release(handle)
|
||||
// when the returned entry is no longer needed. If `handle` is nullptr, it is
|
||||
// as if Release is called immediately after Insert.
|
||||
//
|
||||
// Regardless of whether the item was inserted into the cache,
|
||||
// it will attempt to insert it into the secondary cache if one is
|
||||
// configured, and the helper supports it.
|
||||
// The cache implementation must support a secondary cache, otherwise
|
||||
// the item is only inserted into the primary cache. It may
|
||||
// defer the insertion to the secondary cache as it sees fit.
|
||||
//
|
||||
// When the inserted entry is no longer needed, it will be destroyed using
|
||||
// helper->del_cb (if non-nullptr).
|
||||
virtual Status Insert(const Slice& key, ObjectPtr obj,
|
||||
const CacheItemHelper* helper, size_t charge,
|
||||
Handle** handle = nullptr,
|
||||
Priority priority = Priority::LOW) = 0;
|
||||
|
||||
// Lookup the key, returning nullptr if not found. If found, returns
|
||||
// a handle to the mapping that must eventually be passed to Release().
|
||||
//
|
||||
// If a non-nullptr helper argument is provided with a non-nullptr
|
||||
// create_cb, and a secondary cache is configured, then the secondary
|
||||
// cache is also queried if lookup in the primary cache fails. If found
|
||||
// in secondary cache, the provided create_db and create_context are
|
||||
// used to promote the entry to an object in the primary cache.
|
||||
// In that case, the helper may be saved and used later when the object
|
||||
// is evicted, so as usual, the pointed-to helper must outlive the cache.
|
||||
//
|
||||
// ======================== Async Lookup (wait=false) ======================
|
||||
// When wait=false, the handle returned might be in any of three states:
|
||||
// * Present - If Value() != nullptr, then the result is present and
|
||||
// the handle can be used just as if wait=true.
|
||||
// * Pending, not ready (IsReady() == false) - secondary cache is still
|
||||
// working to retrieve the value. Might become ready any time.
|
||||
// * Pending, ready (IsReady() == true) - secondary cache has the value
|
||||
// but it has not been loaded as an object into primary cache. Call to
|
||||
// Wait()/WaitAll() will not block.
|
||||
//
|
||||
// IMPORTANT: Pending handles are not thread-safe, and only these functions
|
||||
// are allowed on them: Value(), IsReady(), Wait(), WaitAll(). Even Release()
|
||||
// can only come after Wait() or WaitAll() even though a reference is held.
|
||||
//
|
||||
// Only Wait()/WaitAll() gets a Handle out of a Pending state. (Waiting is
|
||||
// safe and has no effect on other handle states.) After waiting on a Handle,
|
||||
// it is in one of two states:
|
||||
// * Present - if Value() != nullptr
|
||||
// * Failed - if Value() == nullptr, such as if the secondary cache
|
||||
// initially thought it had the value but actually did not.
|
||||
//
|
||||
// Note that given an arbitrary Handle, the only way to distinguish the
|
||||
// Pending+ready state from the Failed state is to Wait() on it. A cache
|
||||
// entry not compatible with secondary cache can also have Value()==nullptr
|
||||
// like the Failed state, but this is not generally a concern.
|
||||
virtual Handle* Lookup(const Slice& key,
|
||||
const CacheItemHelper* helper = nullptr,
|
||||
CreateContext* create_context = nullptr,
|
||||
Priority priority = Priority::LOW, bool wait = true,
|
||||
Statistics* stats = nullptr) = 0;
|
||||
|
||||
// Convenience wrapper when secondary cache not supported
|
||||
inline Handle* BasicLookup(const Slice& key, Statistics* stats) {
|
||||
return Lookup(key, nullptr, nullptr, Priority::LOW, true, stats);
|
||||
}
|
||||
|
||||
// Increments the reference count for the handle if it refers to an entry in
|
||||
// the cache. Returns true if refcount was incremented; otherwise, returns
|
||||
// false.
|
||||
// REQUIRES: handle must have been returned by a method on *this.
|
||||
virtual bool Ref(Handle* handle) = 0;
|
||||
|
||||
/**
|
||||
* Release a mapping returned by a previous Lookup(). A released entry might
|
||||
* still remain in cache in case it is later looked up by others. If
|
||||
* erase_if_last_ref is set then it also erases it from the cache if there is
|
||||
* no other reference to it. Erasing it should call the deleter function that
|
||||
* was provided when the entry was inserted.
|
||||
*
|
||||
* Returns true if the entry was also erased.
|
||||
*/
|
||||
// REQUIRES: handle must not have been released yet.
|
||||
// REQUIRES: handle must have been returned by a method on *this.
|
||||
virtual bool Release(Handle* handle, bool erase_if_last_ref = false) = 0;
|
||||
|
||||
// Return the object assiciated with a handle returned by a successful
|
||||
// Lookup(). For historical reasons, this is also known at the "value"
|
||||
// associated with the key.
|
||||
// REQUIRES: handle must not have been released yet.
|
||||
// REQUIRES: handle must have been returned by a method on *this.
|
||||
virtual ObjectPtr Value(Handle* handle) = 0;
|
||||
|
||||
// If the cache contains the entry for the key, erase it. Note that the
|
||||
// underlying entry will be kept around until all existing handles
|
||||
// to it have been released.
|
||||
virtual void Erase(const Slice& key) = 0;
|
||||
// Return a new numeric id. May be used by multiple clients who are
|
||||
// sharding the same cache to partition the key space. Typically the
|
||||
// client will allocate a new id at startup and prepend the id to
|
||||
// its cache keys.
|
||||
virtual uint64_t NewId() = 0;
|
||||
|
||||
// sets the maximum configured capacity of the cache. When the new
|
||||
// capacity is less than the old capacity and the existing usage is
|
||||
// greater than new capacity, the implementation will do its best job to
|
||||
// purge the released entries from the cache in order to lower the usage
|
||||
virtual void SetCapacity(size_t capacity) = 0;
|
||||
|
||||
// Set whether to return error on insertion when cache reaches its full
|
||||
// capacity.
|
||||
virtual void SetStrictCapacityLimit(bool strict_capacity_limit) = 0;
|
||||
|
||||
// Get the flag whether to return error on insertion when cache reaches its
|
||||
// full capacity.
|
||||
virtual bool HasStrictCapacityLimit() const = 0;
|
||||
|
||||
// Returns the maximum configured capacity of the cache
|
||||
virtual size_t GetCapacity() const = 0;
|
||||
|
||||
// Returns the memory size for the entries residing in the cache.
|
||||
virtual size_t GetUsage() const = 0;
|
||||
|
||||
// Returns the number of entries currently tracked in the table. SIZE_MAX
|
||||
// means "not supported." This is used for inspecting the load factor, along
|
||||
// with GetTableAddressCount().
|
||||
virtual size_t GetOccupancyCount() const { return SIZE_MAX; }
|
||||
|
||||
// Returns the number of ways the hash function is divided for addressing
|
||||
// entries. Zero means "not supported." This is used for inspecting the load
|
||||
// factor, along with GetOccupancyCount().
|
||||
virtual size_t GetTableAddressCount() const { return 0; }
|
||||
|
||||
// Returns the memory size for a specific entry in the cache.
|
||||
virtual size_t GetUsage(Handle* handle) const = 0;
|
||||
|
||||
// Returns the memory size for the entries in use by the system
|
||||
virtual size_t GetPinnedUsage() const = 0;
|
||||
|
||||
// Returns the charge for the specific entry in the cache.
|
||||
virtual size_t GetCharge(Handle* handle) const = 0;
|
||||
|
||||
// Returns the helper for the specified entry.
|
||||
virtual const CacheItemHelper* GetCacheItemHelper(Handle* handle) const = 0;
|
||||
|
||||
// Call this on shutdown if you want to speed it up. Cache will disown
|
||||
// any underlying data and will not free it on delete. This call will leak
|
||||
// memory - call this only if you're shutting down the process.
|
||||
// Any attempts of using cache after this call will fail terribly.
|
||||
// Always delete the DB object before calling this method!
|
||||
virtual void DisownData() {
|
||||
// default implementation is noop
|
||||
}
|
||||
|
||||
struct ApplyToAllEntriesOptions {
|
||||
// If the Cache uses locks, setting `average_entries_per_lock` to
|
||||
// a higher value suggests iterating over more entries each time a lock
|
||||
// is acquired, likely reducing the time for ApplyToAllEntries but
|
||||
// increasing latency for concurrent users of the Cache. Setting
|
||||
// `average_entries_per_lock` to a smaller value could be helpful if
|
||||
// callback is relatively expensive, such as using large data structures.
|
||||
size_t average_entries_per_lock = 256;
|
||||
};
|
||||
|
||||
// Apply a callback to all entries in the cache. The Cache must ensure
|
||||
// thread safety but does not guarantee that a consistent snapshot of all
|
||||
// entries is iterated over if other threads are operating on the Cache
|
||||
// also.
|
||||
virtual void ApplyToAllEntries(
|
||||
const std::function<void(const Slice& key, ObjectPtr obj, size_t charge,
|
||||
const CacheItemHelper* helper)>& callback,
|
||||
const ApplyToAllEntriesOptions& opts) = 0;
|
||||
|
||||
// Remove all entries.
|
||||
// Prerequisite: no entry is referenced.
|
||||
virtual void EraseUnRefEntries() = 0;
|
||||
|
||||
virtual std::string GetPrintableOptions() const { return ""; }
|
||||
|
||||
// Check for any warnings or errors in the operation of the cache and
|
||||
// report them to the logger. This is intended only to be called
|
||||
// periodically so does not need to be very efficient. (Obscure calling
|
||||
// conventions for Logger inherited from env.h)
|
||||
virtual void ReportProblems(
|
||||
const std::shared_ptr<Logger>& /*info_log*/) const {}
|
||||
|
||||
MemoryAllocator* memory_allocator() const { return memory_allocator_.get(); }
|
||||
|
||||
// EXPERIMENTAL
|
||||
// The following APIs are experimental and might change in the future.
|
||||
|
||||
// Release a mapping returned by a previous Lookup(). The "useful"
|
||||
// parameter specifies whether the data was actually used or not,
|
||||
// which may be used by the cache implementation to decide whether
|
||||
// to consider it as a hit for retention purposes. As noted elsewhere,
|
||||
// "pending" handles require Wait()/WaitAll() before Release().
|
||||
virtual bool Release(Handle* handle, bool /*useful*/,
|
||||
bool erase_if_last_ref) {
|
||||
return Release(handle, erase_if_last_ref);
|
||||
}
|
||||
|
||||
// Determines if the handle returned by Lookup() can give a value without
|
||||
// blocking, though Wait()/WaitAll() might be required to publish it to
|
||||
// Value(). See secondary cache compatible Lookup() above for details.
|
||||
// This call is not thread safe on "pending" handles.
|
||||
virtual bool IsReady(Handle* /*handle*/) { return true; }
|
||||
|
||||
// Convert a "pending" handle into a full thread-shareable handle by
|
||||
// * If necessary, wait until secondary cache finishes loading the value.
|
||||
// * Construct the object for primary cache and set it in the handle.
|
||||
// Even after Wait() on a pending handle, the caller must check for
|
||||
// Value() == nullptr in case of failure. This call is not thread-safe
|
||||
// on pending handles. This call has no effect on non-pending handles.
|
||||
// See secondary cache compatible Lookup() above for details.
|
||||
virtual void Wait(Handle* /*handle*/) {}
|
||||
|
||||
// Wait for a vector of handles to become ready. As with Wait(), the user
|
||||
// should check the Value() of each handle for nullptr. This call is not
|
||||
// thread-safe on pending handles.
|
||||
virtual void WaitAll(std::vector<Handle*>& /*handles*/) {}
|
||||
|
||||
private:
|
||||
std::shared_ptr<MemoryAllocator> memory_allocator_;
|
||||
};
|
||||
|
||||
// A wrapper around Cache that can easily be extended with instrumentation,
|
||||
// etc.
|
||||
class CacheWrapper : public Cache {
|
||||
public:
|
||||
explicit CacheWrapper(std::shared_ptr<Cache> target)
|
||||
: target_(std::move(target)) {}
|
||||
|
||||
// Only function that derived class must provide
|
||||
// const char* Name() const override { ... }
|
||||
|
||||
Status Insert(const Slice& key, ObjectPtr value,
|
||||
const CacheItemHelper* helper, size_t charge,
|
||||
Handle** handle = nullptr,
|
||||
Priority priority = Priority::LOW) override {
|
||||
return target_->Insert(key, value, helper, charge, handle, priority);
|
||||
}
|
||||
|
||||
Handle* Lookup(const Slice& key, const CacheItemHelper* helper,
|
||||
CreateContext* create_context,
|
||||
Priority priority = Priority::LOW, bool wait = true,
|
||||
Statistics* stats = nullptr) override {
|
||||
return target_->Lookup(key, helper, create_context, priority, wait, stats);
|
||||
}
|
||||
|
||||
bool Ref(Handle* handle) override { return target_->Ref(handle); }
|
||||
|
||||
using Cache::Release;
|
||||
bool Release(Handle* handle, bool erase_if_last_ref = false) override {
|
||||
return target_->Release(handle, erase_if_last_ref);
|
||||
}
|
||||
|
||||
ObjectPtr Value(Handle* handle) override { return target_->Value(handle); }
|
||||
|
||||
void Erase(const Slice& key) override { target_->Erase(key); }
|
||||
uint64_t NewId() override { return target_->NewId(); }
|
||||
|
||||
void SetCapacity(size_t capacity) override { target_->SetCapacity(capacity); }
|
||||
|
||||
void SetStrictCapacityLimit(bool strict_capacity_limit) override {
|
||||
target_->SetStrictCapacityLimit(strict_capacity_limit);
|
||||
}
|
||||
|
||||
bool HasStrictCapacityLimit() const override {
|
||||
return target_->HasStrictCapacityLimit();
|
||||
}
|
||||
|
||||
size_t GetCapacity() const override { return target_->GetCapacity(); }
|
||||
|
||||
size_t GetUsage() const override { return target_->GetUsage(); }
|
||||
|
||||
size_t GetUsage(Handle* handle) const override {
|
||||
return target_->GetUsage(handle);
|
||||
}
|
||||
|
||||
size_t GetPinnedUsage() const override { return target_->GetPinnedUsage(); }
|
||||
|
||||
size_t GetCharge(Handle* handle) const override {
|
||||
return target_->GetCharge(handle);
|
||||
}
|
||||
|
||||
const CacheItemHelper* GetCacheItemHelper(Handle* handle) const override {
|
||||
return target_->GetCacheItemHelper(handle);
|
||||
}
|
||||
|
||||
void ApplyToAllEntries(
|
||||
const std::function<void(const Slice& key, ObjectPtr value, size_t charge,
|
||||
const CacheItemHelper* helper)>& callback,
|
||||
const ApplyToAllEntriesOptions& opts) override {
|
||||
target_->ApplyToAllEntries(callback, opts);
|
||||
}
|
||||
|
||||
void EraseUnRefEntries() override { target_->EraseUnRefEntries(); }
|
||||
|
||||
protected:
|
||||
std::shared_ptr<Cache> target_;
|
||||
};
|
||||
|
||||
// Useful for cache entries requiring no clean-up, such as for cache
|
||||
// reservations
|
||||
inline constexpr Cache::CacheItemHelper kNoopCacheItemHelper{};
|
||||
|
||||
} // namespace ROCKSDB_NAMESPACE
|
|
@ -7,26 +7,21 @@
|
|||
// Use of this source code is governed by a BSD-style license that can be
|
||||
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
||||
//
|
||||
// Various APIs for creating and customizing read caches in RocksDB.
|
||||
// Various APIs for configuring, creating, and monitoring read caches.
|
||||
|
||||
#pragma once
|
||||
|
||||
#include <cstdint>
|
||||
#include <functional>
|
||||
#include <memory>
|
||||
#include <string>
|
||||
|
||||
#include "rocksdb/compression_type.h"
|
||||
#include "rocksdb/memory_allocator.h"
|
||||
#include "rocksdb/slice.h"
|
||||
#include "rocksdb/statistics.h"
|
||||
#include "rocksdb/status.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
||||
class Cache;
|
||||
class Cache; // defined in advanced_cache.h
|
||||
struct ConfigOptions;
|
||||
class Logger;
|
||||
class MemoryAllocator;
|
||||
class SecondaryCache;
|
||||
|
||||
// Classifications of block cache entries.
|
||||
|
@ -352,426 +347,4 @@ extern std::shared_ptr<Cache> NewClockCache(
|
|||
CacheMetadataChargePolicy metadata_charge_policy =
|
||||
kDefaultCacheMetadataChargePolicy);
|
||||
|
||||
// A Cache maps keys to objects resident in memory, tracks reference counts
|
||||
// on those key-object entries, and is able to remove unreferenced entries
|
||||
// whenever it wants. All operations are fully thread safe except as noted.
|
||||
// Inserted entries have a specified "charge" which is some quantity in
|
||||
// unspecified units, typically bytes of memory used. A Cache will typically
|
||||
// have a finite capacity in units of charge, and evict entries as needed
|
||||
// to stay at or below that capacity.
|
||||
//
|
||||
// NOTE: This API is for expert use only and is more intended for providing
|
||||
// custom implementations than for calling into. It is subject to change
|
||||
// as RocksDB evolves, especially the RocksDB block cache.
|
||||
//
|
||||
// INTERNAL: See typed_cache.h for convenient wrappers on top of this API.
|
||||
class Cache {
|
||||
public: // types hidden from API client
|
||||
// Opaque handle to an entry stored in the cache.
|
||||
struct Handle {};
|
||||
|
||||
public: // types hidden from Cache implementation
|
||||
// Pointer to cached object of unspecified type. (This type alias is
|
||||
// provided for clarity, not really for type checking.)
|
||||
using ObjectPtr = void*;
|
||||
|
||||
// Opaque object providing context (settings, etc.) to create objects
|
||||
// for primary cache from saved (serialized) secondary cache entries.
|
||||
struct CreateContext {};
|
||||
|
||||
public: // type defs
|
||||
// Depending on implementation, cache entries with higher priority levels
|
||||
// could be less likely to get evicted than entries with lower priority
|
||||
// levels. The "high" priority level applies to certain SST metablocks (e.g.
|
||||
// index and filter blocks) if the option
|
||||
// cache_index_and_filter_blocks_with_high_priority is set. The "low" priority
|
||||
// level is used for other kinds of SST blocks (most importantly, data
|
||||
// blocks), as well as the above metablocks in case
|
||||
// cache_index_and_filter_blocks_with_high_priority is
|
||||
// not set. The "bottom" priority level is for BlobDB's blob values.
|
||||
enum class Priority { HIGH, LOW, BOTTOM };
|
||||
|
||||
// A set of callbacks to allow objects in the primary block cache to be
|
||||
// be persisted in a secondary cache. The purpose of the secondary cache
|
||||
// is to support other ways of caching the object, such as persistent or
|
||||
// compressed data, that may require the object to be parsed and transformed
|
||||
// in some way. Since the primary cache holds C++ objects and the secondary
|
||||
// cache may only hold flat data that doesn't need relocation, these
|
||||
// callbacks need to be provided by the user of the block
|
||||
// cache to do the conversion.
|
||||
// The CacheItemHelper is passed to Insert() and Lookup(). It has pointers
|
||||
// to callback functions for size, saving and deletion of the
|
||||
// object. The callbacks are defined in C-style in order to make them
|
||||
// stateless and not add to the cache metadata size.
|
||||
// Saving multiple std::function objects will take up 32 bytes per
|
||||
// function, even if its not bound to an object and does no capture.
|
||||
//
|
||||
// All the callbacks are C-style function pointers in order to simplify
|
||||
// lifecycle management. Objects in the cache can outlive the parent DB,
|
||||
// so anything required for these operations should be contained in the
|
||||
// object itself.
|
||||
//
|
||||
// The SizeCallback takes a pointer to the object and returns the size
|
||||
// of the persistable data. It can be used by the secondary cache to allocate
|
||||
// memory if needed.
|
||||
//
|
||||
// RocksDB callbacks are NOT exception-safe. A callback completing with an
|
||||
// exception can lead to undefined behavior in RocksDB, including data loss,
|
||||
// unreported corruption, deadlocks, and more.
|
||||
using SizeCallback = size_t (*)(ObjectPtr obj);
|
||||
|
||||
// The SaveToCallback takes an object pointer and saves the persistable
|
||||
// data into a buffer. The secondary cache may decide to not store it in a
|
||||
// contiguous buffer, in which case this callback will be called multiple
|
||||
// times with increasing offset
|
||||
using SaveToCallback = Status (*)(ObjectPtr from_obj, size_t from_offset,
|
||||
size_t length, char* out_buf);
|
||||
|
||||
// A function pointer type for destruction of a cache object. This will
|
||||
// typically call the destructor for the appropriate type of the object.
|
||||
// The Cache is responsible for copying and reclaiming space for the key,
|
||||
// but objects are managed in part using this callback. Generally a DeleterFn
|
||||
// can be nullptr if the ObjectPtr does not need destruction (e.g. nullptr or
|
||||
// pointer into static data).
|
||||
using DeleterFn = void (*)(ObjectPtr obj, MemoryAllocator* allocator);
|
||||
|
||||
// The CreateCallback is takes in a buffer from the NVM cache and constructs
|
||||
// an object using it. The callback doesn't have ownership of the buffer and
|
||||
// should copy the contents into its own buffer. The CreateContext* is
|
||||
// provided by Lookup and may be used to follow DB- or CF-specific settings.
|
||||
// In case of some error, non-OK is returned and the caller should ignore
|
||||
// any result in out_obj. (The implementation must clean up after itself.)
|
||||
using CreateCallback = Status (*)(const Slice& data, CreateContext* context,
|
||||
MemoryAllocator* allocator,
|
||||
ObjectPtr* out_obj, size_t* out_charge);
|
||||
|
||||
// A struct with pointers to helper functions for spilling items from the
|
||||
// cache into the secondary cache. May be extended in the future. An
|
||||
// instance of this struct is expected to outlive the cache.
|
||||
struct CacheItemHelper {
|
||||
// Function for deleting an object on its removal from the Cache.
|
||||
// nullptr is only for entries that require no destruction, such as
|
||||
// "placeholder" cache entries with nullptr object.
|
||||
DeleterFn del_cb; // (<- Most performance critical)
|
||||
// Next three are used for persisting values as described above.
|
||||
// If any is nullptr, then all three should be nullptr and persisting the
|
||||
// entry to/from secondary cache is not supported.
|
||||
SizeCallback size_cb;
|
||||
SaveToCallback saveto_cb;
|
||||
CreateCallback create_cb;
|
||||
// Classification of the entry for monitoring purposes in block cache.
|
||||
CacheEntryRole role;
|
||||
|
||||
constexpr CacheItemHelper()
|
||||
: del_cb(nullptr),
|
||||
size_cb(nullptr),
|
||||
saveto_cb(nullptr),
|
||||
create_cb(nullptr),
|
||||
role(CacheEntryRole::kMisc) {}
|
||||
|
||||
explicit constexpr CacheItemHelper(CacheEntryRole _role,
|
||||
DeleterFn _del_cb = nullptr,
|
||||
SizeCallback _size_cb = nullptr,
|
||||
SaveToCallback _saveto_cb = nullptr,
|
||||
CreateCallback _create_cb = nullptr)
|
||||
: del_cb(_del_cb),
|
||||
size_cb(_size_cb),
|
||||
saveto_cb(_saveto_cb),
|
||||
create_cb(_create_cb),
|
||||
role(_role) {
|
||||
// Either all three secondary cache callbacks are non-nullptr or
|
||||
// all three are nullptr
|
||||
assert((size_cb != nullptr) == (saveto_cb != nullptr));
|
||||
assert((size_cb != nullptr) == (create_cb != nullptr));
|
||||
}
|
||||
inline bool IsSecondaryCacheCompatible() const {
|
||||
return size_cb != nullptr;
|
||||
}
|
||||
};
|
||||
|
||||
public: // ctor/dtor/create
|
||||
Cache(std::shared_ptr<MemoryAllocator> allocator = nullptr)
|
||||
: memory_allocator_(std::move(allocator)) {}
|
||||
// No copying allowed
|
||||
Cache(const Cache&) = delete;
|
||||
Cache& operator=(const Cache&) = delete;
|
||||
|
||||
// Destroys all remaining entries by calling the associated "deleter"
|
||||
virtual ~Cache() {}
|
||||
|
||||
// Creates a new Cache based on the input value string and returns the result.
|
||||
// Currently, this method can be used to create LRUCaches only
|
||||
// @param config_options
|
||||
// @param value The value might be:
|
||||
// - an old-style cache ("1M") -- equivalent to NewLRUCache(1024*102(
|
||||
// - Name-value option pairs -- "capacity=1M; num_shard_bits=4;
|
||||
// For the LRUCache, the values are defined in LRUCacheOptions.
|
||||
// @param result The new Cache object
|
||||
// @return OK if the cache was successfully created
|
||||
// @return NotFound if an invalid name was specified in the value
|
||||
// @return InvalidArgument if either the options were not valid
|
||||
static Status CreateFromString(const ConfigOptions& config_options,
|
||||
const std::string& value,
|
||||
std::shared_ptr<Cache>* result);
|
||||
|
||||
public: // functions
|
||||
// The type of the Cache
|
||||
virtual const char* Name() const = 0;
|
||||
|
||||
// The Insert and Lookup APIs below are intended to allow cached objects
|
||||
// to be demoted/promoted between the primary block cache and a secondary
|
||||
// cache. The secondary cache could be a non-volatile cache, and will
|
||||
// likely store the object in a different representation. They rely on a
|
||||
// per object CacheItemHelper to do the conversions.
|
||||
// The secondary cache may persist across process and system restarts,
|
||||
// and may even be moved between hosts. Therefore, the cache key must
|
||||
// be repeatable across restarts/reboots, and globally unique if
|
||||
// multiple DBs share the same cache and the set of DBs can change
|
||||
// over time.
|
||||
|
||||
// Insert a mapping from key->object into the cache and assign it
|
||||
// the specified charge against the total cache capacity. If
|
||||
// strict_capacity_limit is true and cache reaches its full capacity,
|
||||
// return Status::MemoryLimit. `obj` must be non-nullptr if compatible
|
||||
// with secondary cache (helper->size_cb != nullptr), because Value() ==
|
||||
// nullptr is reserved for indicating some secondary cache failure cases.
|
||||
// On success, returns OK and takes ownership of `obj`, eventually deleting
|
||||
// it with helper->del_cb. On non-OK return, the caller maintains ownership
|
||||
// of `obj` so will often need to delete it in such cases.
|
||||
//
|
||||
// The helper argument is saved by the cache and will be used when the
|
||||
// inserted object is evicted or considered for promotion to the secondary
|
||||
// cache. Promotion to secondary cache is only enabled if helper->size_cb
|
||||
// != nullptr. The helper must outlive the cache. Callers may use
|
||||
// &kNoopCacheItemHelper as a trivial helper (no deleter for the object,
|
||||
// no secondary cache). `helper` must not be nullptr (efficiency).
|
||||
//
|
||||
// If `handle` is not nullptr and return status is OK, `handle` is set
|
||||
// to a Handle* for the entry. The caller must call this->Release(handle)
|
||||
// when the returned entry is no longer needed. If `handle` is nullptr, it is
|
||||
// as if Release is called immediately after Insert.
|
||||
//
|
||||
// Regardless of whether the item was inserted into the cache,
|
||||
// it will attempt to insert it into the secondary cache if one is
|
||||
// configured, and the helper supports it.
|
||||
// The cache implementation must support a secondary cache, otherwise
|
||||
// the item is only inserted into the primary cache. It may
|
||||
// defer the insertion to the secondary cache as it sees fit.
|
||||
//
|
||||
// When the inserted entry is no longer needed, it will be destroyed using
|
||||
// helper->del_cb (if non-nullptr).
|
||||
virtual Status Insert(const Slice& key, ObjectPtr obj,
|
||||
const CacheItemHelper* helper, size_t charge,
|
||||
Handle** handle = nullptr,
|
||||
Priority priority = Priority::LOW) = 0;
|
||||
|
||||
// Lookup the key, returning nullptr if not found. If found, returns
|
||||
// a handle to the mapping that must eventually be passed to Release().
|
||||
//
|
||||
// If a non-nullptr helper argument is provided with a non-nullptr
|
||||
// create_cb, and a secondary cache is configured, then the secondary
|
||||
// cache is also queried if lookup in the primary cache fails. If found
|
||||
// in secondary cache, the provided create_db and create_context are
|
||||
// used to promote the entry to an object in the primary cache.
|
||||
// In that case, the helper may be saved and used later when the object
|
||||
// is evicted, so as usual, the pointed-to helper must outlive the cache.
|
||||
//
|
||||
// ======================== Async Lookup (wait=false) ======================
|
||||
// When wait=false, the handle returned might be in any of three states:
|
||||
// * Present - If Value() != nullptr, then the result is present and
|
||||
// the handle can be used just as if wait=true.
|
||||
// * Pending, not ready (IsReady() == false) - secondary cache is still
|
||||
// working to retrieve the value. Might become ready any time.
|
||||
// * Pending, ready (IsReady() == true) - secondary cache has the value
|
||||
// but it has not been loaded as an object into primary cache. Call to
|
||||
// Wait()/WaitAll() will not block.
|
||||
//
|
||||
// IMPORTANT: Pending handles are not thread-safe, and only these functions
|
||||
// are allowed on them: Value(), IsReady(), Wait(), WaitAll(). Even Release()
|
||||
// can only come after Wait() or WaitAll() even though a reference is held.
|
||||
//
|
||||
// Only Wait()/WaitAll() gets a Handle out of a Pending state. (Waiting is
|
||||
// safe and has no effect on other handle states.) After waiting on a Handle,
|
||||
// it is in one of two states:
|
||||
// * Present - if Value() != nullptr
|
||||
// * Failed - if Value() == nullptr, such as if the secondary cache
|
||||
// initially thought it had the value but actually did not.
|
||||
//
|
||||
// Note that given an arbitrary Handle, the only way to distinguish the
|
||||
// Pending+ready state from the Failed state is to Wait() on it. A cache
|
||||
// entry not compatible with secondary cache can also have Value()==nullptr
|
||||
// like the Failed state, but this is not generally a concern.
|
||||
virtual Handle* Lookup(const Slice& key,
|
||||
const CacheItemHelper* helper = nullptr,
|
||||
CreateContext* create_context = nullptr,
|
||||
Priority priority = Priority::LOW, bool wait = true,
|
||||
Statistics* stats = nullptr) = 0;
|
||||
|
||||
// Convenience wrapper when secondary cache not supported
|
||||
inline Handle* BasicLookup(const Slice& key, Statistics* stats) {
|
||||
return Lookup(key, nullptr, nullptr, Priority::LOW, true, stats);
|
||||
}
|
||||
|
||||
// Increments the reference count for the handle if it refers to an entry in
|
||||
// the cache. Returns true if refcount was incremented; otherwise, returns
|
||||
// false.
|
||||
// REQUIRES: handle must have been returned by a method on *this.
|
||||
virtual bool Ref(Handle* handle) = 0;
|
||||
|
||||
/**
|
||||
* Release a mapping returned by a previous Lookup(). A released entry might
|
||||
* still remain in cache in case it is later looked up by others. If
|
||||
* erase_if_last_ref is set then it also erases it from the cache if there is
|
||||
* no other reference to it. Erasing it should call the deleter function that
|
||||
* was provided when the entry was inserted.
|
||||
*
|
||||
* Returns true if the entry was also erased.
|
||||
*/
|
||||
// REQUIRES: handle must not have been released yet.
|
||||
// REQUIRES: handle must have been returned by a method on *this.
|
||||
virtual bool Release(Handle* handle, bool erase_if_last_ref = false) = 0;
|
||||
|
||||
// Return the object assiciated with a handle returned by a successful
|
||||
// Lookup(). For historical reasons, this is also known at the "value"
|
||||
// associated with the key.
|
||||
// REQUIRES: handle must not have been released yet.
|
||||
// REQUIRES: handle must have been returned by a method on *this.
|
||||
virtual ObjectPtr Value(Handle* handle) = 0;
|
||||
|
||||
// If the cache contains the entry for the key, erase it. Note that the
|
||||
// underlying entry will be kept around until all existing handles
|
||||
// to it have been released.
|
||||
virtual void Erase(const Slice& key) = 0;
|
||||
// Return a new numeric id. May be used by multiple clients who are
|
||||
// sharding the same cache to partition the key space. Typically the
|
||||
// client will allocate a new id at startup and prepend the id to
|
||||
// its cache keys.
|
||||
virtual uint64_t NewId() = 0;
|
||||
|
||||
// sets the maximum configured capacity of the cache. When the new
|
||||
// capacity is less than the old capacity and the existing usage is
|
||||
// greater than new capacity, the implementation will do its best job to
|
||||
// purge the released entries from the cache in order to lower the usage
|
||||
virtual void SetCapacity(size_t capacity) = 0;
|
||||
|
||||
// Set whether to return error on insertion when cache reaches its full
|
||||
// capacity.
|
||||
virtual void SetStrictCapacityLimit(bool strict_capacity_limit) = 0;
|
||||
|
||||
// Get the flag whether to return error on insertion when cache reaches its
|
||||
// full capacity.
|
||||
virtual bool HasStrictCapacityLimit() const = 0;
|
||||
|
||||
// Returns the maximum configured capacity of the cache
|
||||
virtual size_t GetCapacity() const = 0;
|
||||
|
||||
// Returns the memory size for the entries residing in the cache.
|
||||
virtual size_t GetUsage() const = 0;
|
||||
|
||||
// Returns the number of entries currently tracked in the table. SIZE_MAX
|
||||
// means "not supported." This is used for inspecting the load factor, along
|
||||
// with GetTableAddressCount().
|
||||
virtual size_t GetOccupancyCount() const { return SIZE_MAX; }
|
||||
|
||||
// Returns the number of ways the hash function is divided for addressing
|
||||
// entries. Zero means "not supported." This is used for inspecting the load
|
||||
// factor, along with GetOccupancyCount().
|
||||
virtual size_t GetTableAddressCount() const { return 0; }
|
||||
|
||||
// Returns the memory size for a specific entry in the cache.
|
||||
virtual size_t GetUsage(Handle* handle) const = 0;
|
||||
|
||||
// Returns the memory size for the entries in use by the system
|
||||
virtual size_t GetPinnedUsage() const = 0;
|
||||
|
||||
// Returns the charge for the specific entry in the cache.
|
||||
virtual size_t GetCharge(Handle* handle) const = 0;
|
||||
|
||||
// Returns the helper for the specified entry.
|
||||
virtual const CacheItemHelper* GetCacheItemHelper(Handle* handle) const = 0;
|
||||
|
||||
// Call this on shutdown if you want to speed it up. Cache will disown
|
||||
// any underlying data and will not free it on delete. This call will leak
|
||||
// memory - call this only if you're shutting down the process.
|
||||
// Any attempts of using cache after this call will fail terribly.
|
||||
// Always delete the DB object before calling this method!
|
||||
virtual void DisownData() {
|
||||
// default implementation is noop
|
||||
}
|
||||
|
||||
struct ApplyToAllEntriesOptions {
|
||||
// If the Cache uses locks, setting `average_entries_per_lock` to
|
||||
// a higher value suggests iterating over more entries each time a lock
|
||||
// is acquired, likely reducing the time for ApplyToAllEntries but
|
||||
// increasing latency for concurrent users of the Cache. Setting
|
||||
// `average_entries_per_lock` to a smaller value could be helpful if
|
||||
// callback is relatively expensive, such as using large data structures.
|
||||
size_t average_entries_per_lock = 256;
|
||||
};
|
||||
|
||||
// Apply a callback to all entries in the cache. The Cache must ensure
|
||||
// thread safety but does not guarantee that a consistent snapshot of all
|
||||
// entries is iterated over if other threads are operating on the Cache
|
||||
// also.
|
||||
virtual void ApplyToAllEntries(
|
||||
const std::function<void(const Slice& key, ObjectPtr obj, size_t charge,
|
||||
const CacheItemHelper* helper)>& callback,
|
||||
const ApplyToAllEntriesOptions& opts) = 0;
|
||||
|
||||
// Remove all entries.
|
||||
// Prerequisite: no entry is referenced.
|
||||
virtual void EraseUnRefEntries() = 0;
|
||||
|
||||
virtual std::string GetPrintableOptions() const { return ""; }
|
||||
|
||||
// Check for any warnings or errors in the operation of the cache and
|
||||
// report them to the logger. This is intended only to be called
|
||||
// periodically so does not need to be very efficient. (Obscure calling
|
||||
// conventions for Logger inherited from env.h)
|
||||
virtual void ReportProblems(
|
||||
const std::shared_ptr<Logger>& /*info_log*/) const {}
|
||||
|
||||
MemoryAllocator* memory_allocator() const { return memory_allocator_.get(); }
|
||||
|
||||
// EXPERIMENTAL
|
||||
// The following APIs are experimental and might change in the future.
|
||||
|
||||
// Release a mapping returned by a previous Lookup(). The "useful"
|
||||
// parameter specifies whether the data was actually used or not,
|
||||
// which may be used by the cache implementation to decide whether
|
||||
// to consider it as a hit for retention purposes. As noted elsewhere,
|
||||
// "pending" handles require Wait()/WaitAll() before Release().
|
||||
virtual bool Release(Handle* handle, bool /*useful*/,
|
||||
bool erase_if_last_ref) {
|
||||
return Release(handle, erase_if_last_ref);
|
||||
}
|
||||
|
||||
// Determines if the handle returned by Lookup() can give a value without
|
||||
// blocking, though Wait()/WaitAll() might be required to publish it to
|
||||
// Value(). See secondary cache compatible Lookup() above for details.
|
||||
// This call is not thread safe on "pending" handles.
|
||||
virtual bool IsReady(Handle* /*handle*/) { return true; }
|
||||
|
||||
// Convert a "pending" handle into a full thread-shareable handle by
|
||||
// * If necessary, wait until secondary cache finishes loading the value.
|
||||
// * Construct the object for primary cache and set it in the handle.
|
||||
// Even after Wait() on a pending handle, the caller must check for
|
||||
// Value() == nullptr in case of failure. This call is not thread-safe
|
||||
// on pending handles. This call has no effect on non-pending handles.
|
||||
// See secondary cache compatible Lookup() above for details.
|
||||
virtual void Wait(Handle* /*handle*/) {}
|
||||
|
||||
// Wait for a vector of handles to become ready. As with Wait(), the user
|
||||
// should check the Value() of each handle for nullptr. This call is not
|
||||
// thread-safe on pending handles.
|
||||
virtual void WaitAll(std::vector<Handle*>& /*handles*/) {}
|
||||
|
||||
private:
|
||||
std::shared_ptr<MemoryAllocator> memory_allocator_;
|
||||
};
|
||||
|
||||
// Useful for cache entries requiring no clean-up, such as for cache
|
||||
// reservations
|
||||
inline constexpr Cache::CacheItemHelper kNoopCacheItemHelper{};
|
||||
|
||||
} // namespace ROCKSDB_NAMESPACE
|
||||
|
|
|
@ -9,7 +9,7 @@
|
|||
#include <memory>
|
||||
#include <string>
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/customizable.h"
|
||||
#include "rocksdb/slice.h"
|
||||
#include "rocksdb/statistics.h"
|
||||
|
|
|
@ -10,7 +10,7 @@
|
|||
#include <memory>
|
||||
#include <string>
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/env.h"
|
||||
#include "rocksdb/slice.h"
|
||||
#include "rocksdb/statistics.h"
|
||||
|
@ -42,13 +42,10 @@ extern std::shared_ptr<SimCache> NewSimCache(std::shared_ptr<Cache> sim_cache,
|
|||
std::shared_ptr<Cache> cache,
|
||||
int num_shard_bits);
|
||||
|
||||
class SimCache : public Cache {
|
||||
// An abstract base class (public interface) to the SimCache implementation
|
||||
class SimCache : public CacheWrapper {
|
||||
public:
|
||||
SimCache() {}
|
||||
|
||||
~SimCache() override {}
|
||||
|
||||
const char* Name() const override { return "SimCache"; }
|
||||
using CacheWrapper::CacheWrapper;
|
||||
|
||||
// returns the maximum configured capacity of the simcache for simulation
|
||||
virtual size_t GetSimCapacity() const = 0;
|
||||
|
|
|
@ -6,11 +6,10 @@
|
|||
// This file implements the "bridge" between Java and C++ for
|
||||
// ROCKSDB_NAMESPACE::Cache.
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
|
||||
#include <jni.h>
|
||||
|
||||
#include "include/org_rocksdb_Cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
|
||||
/*
|
||||
* Class: org_rocksdb_Cache
|
||||
|
|
|
@ -7,6 +7,8 @@
|
|||
|
||||
#include <jni.h>
|
||||
|
||||
#include <cassert>
|
||||
|
||||
#include "include/org_rocksdb_WriteBufferManager.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksjni/cplusplus_to_java_convert.h"
|
||||
|
|
|
@ -9,6 +9,7 @@
|
|||
|
||||
#include "rocksdb/write_buffer_manager.h"
|
||||
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "test_util/testharness.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
|
|
@ -16,6 +16,7 @@
|
|||
#include "options/options_helper.h"
|
||||
#include "options/options_parser.h"
|
||||
#include "port/port.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/compaction_filter.h"
|
||||
#include "rocksdb/concurrent_task_limiter.h"
|
||||
#include "rocksdb/configurable.h"
|
||||
|
|
|
@ -12,6 +12,7 @@
|
|||
#include "options/options_helper.h"
|
||||
#include "options/options_parser.h"
|
||||
#include "port/port.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/configurable.h"
|
||||
#include "rocksdb/env.h"
|
||||
#include "rocksdb/file_system.h"
|
||||
|
|
|
@ -16,6 +16,7 @@
|
|||
|
||||
#include "db/pinned_iterators_manager.h"
|
||||
#include "port/malloc.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/iterator.h"
|
||||
#include "rocksdb/options.h"
|
||||
#include "rocksdb/statistics.h"
|
||||
|
|
|
@ -13,7 +13,7 @@
|
|||
#include <type_traits>
|
||||
|
||||
#include "port/likely.h"
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/cleanable.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
|
|
@ -8,7 +8,7 @@
|
|||
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
||||
#pragma once
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/table.h"
|
||||
|
||||
namespace ROCKSDB_NAMESPACE {
|
||||
|
|
|
@ -9,7 +9,7 @@
|
|||
|
||||
#include "rocksdb/utilities/leveldb_options.h"
|
||||
|
||||
#include "rocksdb/cache.h"
|
||||
#include "rocksdb/advanced_cache.h"
|
||||
#include "rocksdb/comparator.h"
|
||||
#include "rocksdb/env.h"
|
||||
#include "rocksdb/filter_policy.h"
|
||||
|
|
|
@ -151,20 +151,22 @@ class SimCacheImpl : public SimCache {
|
|||
// capacity for real cache (ShardedLRUCache)
|
||||
// test_capacity for key only cache
|
||||
SimCacheImpl(std::shared_ptr<Cache> sim_cache, std::shared_ptr<Cache> cache)
|
||||
: cache_(cache),
|
||||
: SimCache(cache),
|
||||
key_only_cache_(sim_cache),
|
||||
miss_times_(0),
|
||||
hit_times_(0),
|
||||
stats_(nullptr) {}
|
||||
|
||||
~SimCacheImpl() override {}
|
||||
void SetCapacity(size_t capacity) override { cache_->SetCapacity(capacity); }
|
||||
|
||||
const char* Name() const override { return "SimCache"; }
|
||||
|
||||
void SetCapacity(size_t capacity) override { target_->SetCapacity(capacity); }
|
||||
|
||||
void SetStrictCapacityLimit(bool strict_capacity_limit) override {
|
||||
cache_->SetStrictCapacityLimit(strict_capacity_limit);
|
||||
target_->SetStrictCapacityLimit(strict_capacity_limit);
|
||||
}
|
||||
|
||||
using Cache::Insert;
|
||||
Status Insert(const Slice& key, Cache::ObjectPtr value,
|
||||
const CacheItemHelper* helper, size_t charge, Handle** handle,
|
||||
Priority priority) override {
|
||||
|
@ -184,10 +186,10 @@ class SimCacheImpl : public SimCache {
|
|||
}
|
||||
|
||||
cache_activity_logger_.ReportAdd(key, charge);
|
||||
if (!cache_) {
|
||||
if (!target_) {
|
||||
return Status::OK();
|
||||
}
|
||||
return cache_->Insert(key, value, helper, charge, handle, priority);
|
||||
return target_->Insert(key, value, helper, charge, handle, priority);
|
||||
}
|
||||
|
||||
Handle* Lookup(const Slice& key, const CacheItemHelper* helper,
|
||||
|
@ -195,54 +197,54 @@ class SimCacheImpl : public SimCache {
|
|||
Priority priority = Priority::LOW, bool wait = true,
|
||||
Statistics* stats = nullptr) override {
|
||||
HandleLookup(key, stats);
|
||||
if (!cache_) {
|
||||
if (!target_) {
|
||||
return nullptr;
|
||||
}
|
||||
return cache_->Lookup(key, helper, create_context, priority, wait, stats);
|
||||
return target_->Lookup(key, helper, create_context, priority, wait, stats);
|
||||
}
|
||||
|
||||
bool Ref(Handle* handle) override { return cache_->Ref(handle); }
|
||||
bool Ref(Handle* handle) override { return target_->Ref(handle); }
|
||||
|
||||
using Cache::Release;
|
||||
bool Release(Handle* handle, bool erase_if_last_ref = false) override {
|
||||
return cache_->Release(handle, erase_if_last_ref);
|
||||
return target_->Release(handle, erase_if_last_ref);
|
||||
}
|
||||
|
||||
void Erase(const Slice& key) override {
|
||||
cache_->Erase(key);
|
||||
target_->Erase(key);
|
||||
key_only_cache_->Erase(key);
|
||||
}
|
||||
|
||||
Cache::ObjectPtr Value(Handle* handle) override {
|
||||
return cache_->Value(handle);
|
||||
return target_->Value(handle);
|
||||
}
|
||||
|
||||
uint64_t NewId() override { return cache_->NewId(); }
|
||||
uint64_t NewId() override { return target_->NewId(); }
|
||||
|
||||
size_t GetCapacity() const override { return cache_->GetCapacity(); }
|
||||
size_t GetCapacity() const override { return target_->GetCapacity(); }
|
||||
|
||||
bool HasStrictCapacityLimit() const override {
|
||||
return cache_->HasStrictCapacityLimit();
|
||||
return target_->HasStrictCapacityLimit();
|
||||
}
|
||||
|
||||
size_t GetUsage() const override { return cache_->GetUsage(); }
|
||||
size_t GetUsage() const override { return target_->GetUsage(); }
|
||||
|
||||
size_t GetUsage(Handle* handle) const override {
|
||||
return cache_->GetUsage(handle);
|
||||
return target_->GetUsage(handle);
|
||||
}
|
||||
|
||||
size_t GetCharge(Handle* handle) const override {
|
||||
return cache_->GetCharge(handle);
|
||||
return target_->GetCharge(handle);
|
||||
}
|
||||
|
||||
const CacheItemHelper* GetCacheItemHelper(Handle* handle) const override {
|
||||
return cache_->GetCacheItemHelper(handle);
|
||||
return target_->GetCacheItemHelper(handle);
|
||||
}
|
||||
|
||||
size_t GetPinnedUsage() const override { return cache_->GetPinnedUsage(); }
|
||||
size_t GetPinnedUsage() const override { return target_->GetPinnedUsage(); }
|
||||
|
||||
void DisownData() override {
|
||||
cache_->DisownData();
|
||||
target_->DisownData();
|
||||
key_only_cache_->DisownData();
|
||||
}
|
||||
|
||||
|
@ -250,11 +252,11 @@ class SimCacheImpl : public SimCache {
|
|||
const std::function<void(const Slice& key, ObjectPtr value, size_t charge,
|
||||
const CacheItemHelper* helper)>& callback,
|
||||
const ApplyToAllEntriesOptions& opts) override {
|
||||
cache_->ApplyToAllEntries(callback, opts);
|
||||
target_->ApplyToAllEntries(callback, opts);
|
||||
}
|
||||
|
||||
void EraseUnRefEntries() override {
|
||||
cache_->EraseUnRefEntries();
|
||||
target_->EraseUnRefEntries();
|
||||
key_only_cache_->EraseUnRefEntries();
|
||||
}
|
||||
|
||||
|
@ -295,7 +297,7 @@ class SimCacheImpl : public SimCache {
|
|||
std::string GetPrintableOptions() const override {
|
||||
std::ostringstream oss;
|
||||
oss << " cache_options:" << std::endl;
|
||||
oss << cache_->GetPrintableOptions();
|
||||
oss << target_->GetPrintableOptions();
|
||||
oss << " sim_cache_options:" << std::endl;
|
||||
oss << key_only_cache_->GetPrintableOptions();
|
||||
return oss.str();
|
||||
|
@ -314,7 +316,6 @@ class SimCacheImpl : public SimCache {
|
|||
}
|
||||
|
||||
private:
|
||||
std::shared_ptr<Cache> cache_;
|
||||
std::shared_ptr<Cache> key_only_cache_;
|
||||
std::atomic<uint64_t> miss_times_;
|
||||
std::atomic<uint64_t> hit_times_;
|
||||
|
|
Loading…
Reference in New Issue