2016-02-09 23:12:00 +00:00
|
|
|
// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
|
2013-10-16 21:59:46 +00:00
|
|
|
// This source code is licensed under the BSD-style license found in the
|
|
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
|
|
//
|
2011-03-18 22:37:00 +00:00
|
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
|
|
|
|
#include "db/table_cache.h"
|
|
|
|
|
2015-06-23 17:25:45 +00:00
|
|
|
#include "db/dbformat.h"
|
2011-03-18 22:37:00 +00:00
|
|
|
#include "db/filename.h"
|
2014-01-07 04:29:17 +00:00
|
|
|
#include "db/version_edit.h"
|
2013-02-25 21:58:34 +00:00
|
|
|
|
2013-08-23 15:38:13 +00:00
|
|
|
#include "rocksdb/statistics.h"
|
2015-10-12 22:06:38 +00:00
|
|
|
#include "table/internal_iterator.h"
|
In DB::NewIterator(), try to allocate the whole iterator tree in an arena
Summary:
In this patch, try to allocate the whole iterator tree starting from DBIter from an arena
1. ArenaWrappedDBIter is created when serves as the entry point of an iterator tree, with an arena in it.
2. Add an option to create iterator from arena for following iterators: DBIter, MergingIterator, MemtableIterator, all mem table's iterators, all table reader's iterators and two level iterator.
3. MergeIteratorBuilder is created to incrementally build the tree of internal iterators. It is passed to mem table list and version set and add iterators to it.
Limitations:
(1) Only DB::NewIterator() without tailing uses the arena. Other cases, including readonly DB and compactions are still from malloc
(2) Two level iterator itself is allocated in arena, but not iterators inside it.
Test Plan: make all check
Reviewers: ljin, haobo
Reviewed By: haobo
Subscribers: leveldb, dhruba, yhchiang, igor
Differential Revision: https://reviews.facebook.net/D18513
2014-06-02 23:38:00 +00:00
|
|
|
#include "table/iterator_wrapper.h"
|
2015-09-11 18:36:33 +00:00
|
|
|
#include "table/table_builder.h"
|
2014-01-28 05:58:46 +00:00
|
|
|
#include "table/table_reader.h"
|
2014-09-29 18:09:09 +00:00
|
|
|
#include "table/get_context.h"
|
2011-03-18 22:37:00 +00:00
|
|
|
#include "util/coding.h"
|
Move rate_limiter, write buffering, most perf context instrumentation and most random kill out of Env
Summary: We want to keep Env a think layer for better portability. Less platform dependent codes should be moved out of Env. In this patch, I create a wrapper of file readers and writers, and put rate limiting, write buffering, as well as most perf context instrumentation and random kill out of Env. It will make it easier to maintain multiple Env in the future.
Test Plan: Run all existing unit tests.
Reviewers: anthony, kradhakrishnan, IslamAbdelRahman, yhchiang, igor
Reviewed By: igor
Subscribers: leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D42321
2015-07-17 23:16:11 +00:00
|
|
|
#include "util/file_reader_writer.h"
|
2015-07-08 23:34:48 +00:00
|
|
|
#include "util/perf_context_imp.h"
|
2013-06-07 17:02:28 +00:00
|
|
|
#include "util/stop_watch.h"
|
2015-08-21 05:33:25 +00:00
|
|
|
#include "util/sync_point.h"
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2013-10-04 04:49:15 +00:00
|
|
|
namespace rocksdb {
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2015-06-23 17:25:45 +00:00
|
|
|
namespace {
|
|
|
|
|
|
|
|
template <class T>
|
2011-03-18 22:37:00 +00:00
|
|
|
static void DeleteEntry(const Slice& key, void* value) {
|
2015-06-23 17:25:45 +00:00
|
|
|
T* typed_value = reinterpret_cast<T*>(value);
|
|
|
|
delete typed_value;
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void UnrefEntry(void* arg1, void* arg2) {
|
|
|
|
Cache* cache = reinterpret_cast<Cache*>(arg1);
|
|
|
|
Cache::Handle* h = reinterpret_cast<Cache::Handle*>(arg2);
|
|
|
|
cache->Release(h);
|
|
|
|
}
|
|
|
|
|
2015-08-21 05:33:25 +00:00
|
|
|
static void DeleteTableReader(void* arg1, void* arg2) {
|
|
|
|
TableReader* table_reader = reinterpret_cast<TableReader*>(arg1);
|
|
|
|
delete table_reader;
|
|
|
|
}
|
|
|
|
|
2014-06-13 22:54:19 +00:00
|
|
|
static Slice GetSliceForFileNumber(const uint64_t* file_number) {
|
2014-01-02 18:29:48 +00:00
|
|
|
return Slice(reinterpret_cast<const char*>(file_number),
|
|
|
|
sizeof(*file_number));
|
2013-12-27 00:25:45 +00:00
|
|
|
}
|
|
|
|
|
2015-06-23 17:25:45 +00:00
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
|
|
|
|
void AppendVarint64(IterKey* key, uint64_t v) {
|
|
|
|
char buf[10];
|
|
|
|
auto ptr = EncodeVarint64(buf, v);
|
|
|
|
key->TrimAppend(key->Size(), buf, ptr - buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif // ROCKSDB_LITE
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
2014-09-04 23:18:36 +00:00
|
|
|
TableCache::TableCache(const ImmutableCFOptions& ioptions,
|
|
|
|
const EnvOptions& env_options, Cache* const cache)
|
2015-06-23 17:25:45 +00:00
|
|
|
: ioptions_(ioptions), env_options_(env_options), cache_(cache) {
|
|
|
|
if (ioptions_.row_cache) {
|
|
|
|
// If the same cache is shared by multiple instances, we need to
|
|
|
|
// disambiguate its entries.
|
|
|
|
PutVarint64(&row_cache_id_, ioptions_.row_cache->NewId());
|
|
|
|
}
|
|
|
|
}
|
2011-03-18 22:37:00 +00:00
|
|
|
|
|
|
|
TableCache::~TableCache() {
|
|
|
|
}
|
|
|
|
|
2014-01-07 04:29:17 +00:00
|
|
|
TableReader* TableCache::GetTableReaderFromHandle(Cache::Handle* handle) {
|
|
|
|
return reinterpret_cast<TableReader*>(cache_->Value(handle));
|
|
|
|
}
|
|
|
|
|
|
|
|
void TableCache::ReleaseHandle(Cache::Handle* handle) {
|
|
|
|
cache_->Release(handle);
|
|
|
|
}
|
|
|
|
|
2015-08-21 05:33:25 +00:00
|
|
|
Status TableCache::GetTableReader(
|
|
|
|
const EnvOptions& env_options,
|
|
|
|
const InternalKeyComparator& internal_comparator, const FileDescriptor& fd,
|
2015-08-26 22:25:59 +00:00
|
|
|
bool sequential_mode, bool record_read_stats, HistogramImpl* file_read_hist,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
unique_ptr<TableReader>* table_reader, bool skip_filters, int level) {
|
2015-08-21 05:33:25 +00:00
|
|
|
std::string fname =
|
|
|
|
TableFileName(ioptions_.db_paths, fd.GetNumber(), fd.GetPathId());
|
|
|
|
unique_ptr<RandomAccessFile> file;
|
|
|
|
Status s = ioptions_.env->NewRandomAccessFile(fname, &file, env_options);
|
2015-08-26 22:25:59 +00:00
|
|
|
if (sequential_mode && ioptions_.compaction_readahead_size > 0) {
|
|
|
|
file = NewReadaheadRandomAccessFile(std::move(file),
|
|
|
|
ioptions_.compaction_readahead_size);
|
|
|
|
}
|
2015-08-21 05:33:25 +00:00
|
|
|
RecordTick(ioptions_.statistics, NO_FILE_OPENS);
|
|
|
|
if (s.ok()) {
|
2015-08-26 22:25:59 +00:00
|
|
|
if (!sequential_mode && ioptions_.advise_random_on_open) {
|
2015-08-21 05:33:25 +00:00
|
|
|
file->Hint(RandomAccessFile::RANDOM);
|
|
|
|
}
|
|
|
|
StopWatch sw(ioptions_.env, ioptions_.statistics, TABLE_OPEN_IO_MICROS);
|
|
|
|
std::unique_ptr<RandomAccessFileReader> file_reader(
|
|
|
|
new RandomAccessFileReader(std::move(file), ioptions_.env,
|
|
|
|
ioptions_.statistics, record_read_stats,
|
|
|
|
file_read_hist));
|
|
|
|
s = ioptions_.table_factory->NewTableReader(
|
Skip bottom-level filter block caching when hit-optimized
Summary:
When Get() or NewIterator() trigger file loads, skip caching the filter block if
(1) optimize_filters_for_hits is set and (2) the file is on the bottommost
level. Also skip checking filters under the same conditions, which means that
for a preloaded file or a file that was trivially-moved to the bottom level, its
filter block will eventually expire from the cache.
- added parameters/instance variables in various places in order to propagate the config ("skip_filters") from version_set to block_based_table_reader
- in BlockBasedTable::Rep, this optimization prevents filter from being loaded when the file is opened simply by setting filter_policy = nullptr
- in BlockBasedTable::Get/BlockBasedTable::NewIterator, this optimization prevents filter from being used (even if it was loaded already) by setting filter = nullptr
Test Plan:
updated unit test:
$ ./db_test --gtest_filter=DBTest.OptimizeFiltersForHits
will also run 'make check'
Reviewers: sdong, igor, paultuckfield, anthony, rven, kradhakrishnan, IslamAbdelRahman, yhchiang
Reviewed By: yhchiang
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D51633
2015-12-23 18:15:07 +00:00
|
|
|
TableReaderOptions(ioptions_, env_options, internal_comparator,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
skip_filters, level),
|
2015-09-11 18:36:33 +00:00
|
|
|
std::move(file_reader), fd.GetFileSize(), table_reader);
|
2015-08-21 05:33:25 +00:00
|
|
|
TEST_SYNC_POINT("TableCache::GetTableReader:0");
|
|
|
|
}
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2014-09-04 23:18:36 +00:00
|
|
|
Status TableCache::FindTable(const EnvOptions& env_options,
|
2014-01-27 21:53:22 +00:00
|
|
|
const InternalKeyComparator& internal_comparator,
|
2014-06-13 22:54:19 +00:00
|
|
|
const FileDescriptor& fd, Cache::Handle** handle,
|
Measure file read latency histogram per level
Summary: In internal stats, remember read latency histogram, if statistics is enabled. It can be retrieved from DB::GetProperty() with "rocksdb.dbstats" property, if it is enabled.
Test Plan: Manually run db_bench and prints out "rocksdb.dbstats" by hand and make sure it prints out as expected
Reviewers: igor, IslamAbdelRahman, rven, kradhakrishnan, anthony, yhchiang
Reviewed By: yhchiang
Subscribers: MarkCallaghan, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D44193
2015-08-13 21:35:54 +00:00
|
|
|
const bool no_io, bool record_read_stats,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
HistogramImpl* file_read_hist, bool skip_filters,
|
|
|
|
int level) {
|
2015-07-08 23:34:48 +00:00
|
|
|
PERF_TIMER_GUARD(find_table_nanos);
|
2012-04-17 15:36:46 +00:00
|
|
|
Status s;
|
2014-07-02 16:54:20 +00:00
|
|
|
uint64_t number = fd.GetNumber();
|
|
|
|
Slice key = GetSliceForFileNumber(&number);
|
2012-04-17 15:36:46 +00:00
|
|
|
*handle = cache_->Lookup(key);
|
2015-08-21 05:33:25 +00:00
|
|
|
TEST_SYNC_POINT_CALLBACK("TableCache::FindTable:0",
|
|
|
|
const_cast<bool*>(&no_io));
|
|
|
|
|
2013-02-25 21:58:34 +00:00
|
|
|
if (*handle == nullptr) {
|
2015-08-26 17:10:26 +00:00
|
|
|
if (no_io) { // Don't do IO and return a not-found status
|
2013-08-25 05:48:51 +00:00
|
|
|
return Status::Incomplete("Table not found in table_cache, no_io is set");
|
2013-07-12 23:56:52 +00:00
|
|
|
}
|
2013-10-30 17:52:33 +00:00
|
|
|
unique_ptr<TableReader> table_reader;
|
2015-08-21 05:33:25 +00:00
|
|
|
s = GetTableReader(env_options, internal_comparator, fd,
|
2015-08-26 22:25:59 +00:00
|
|
|
false /* sequential mode */, record_read_stats,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
file_read_hist, &table_reader, skip_filters, level);
|
2011-03-18 22:37:00 +00:00
|
|
|
if (!s.ok()) {
|
2013-10-30 17:52:33 +00:00
|
|
|
assert(table_reader == nullptr);
|
2014-09-04 23:18:36 +00:00
|
|
|
RecordTick(ioptions_.statistics, NO_FILE_ERRORS);
|
2011-03-18 22:37:00 +00:00
|
|
|
// We do not cache error results so that if the error is transient,
|
|
|
|
// or somebody repairs the file, we recover automatically.
|
2012-04-17 15:36:46 +00:00
|
|
|
} else {
|
2016-03-11 01:35:19 +00:00
|
|
|
s = cache_->Insert(key, table_reader.get(), 1, &DeleteEntry<TableReader>,
|
|
|
|
handle);
|
|
|
|
if (s.ok()) {
|
|
|
|
// Release ownership of table reader.
|
|
|
|
table_reader.release();
|
|
|
|
}
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
2012-04-17 15:36:46 +00:00
|
|
|
}
|
|
|
|
return s;
|
|
|
|
}
|
2011-03-18 22:37:00 +00:00
|
|
|
|
2015-10-12 22:06:38 +00:00
|
|
|
InternalIterator* TableCache::NewIterator(
|
|
|
|
const ReadOptions& options, const EnvOptions& env_options,
|
|
|
|
const InternalKeyComparator& icomparator, const FileDescriptor& fd,
|
|
|
|
TableReader** table_reader_ptr, HistogramImpl* file_read_hist,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
bool for_compaction, Arena* arena, bool skip_filters, int level) {
|
2015-07-08 23:34:48 +00:00
|
|
|
PERF_TIMER_GUARD(new_table_iterator_nanos);
|
|
|
|
|
2013-10-30 17:52:33 +00:00
|
|
|
if (table_reader_ptr != nullptr) {
|
|
|
|
*table_reader_ptr = nullptr;
|
2012-04-17 15:36:46 +00:00
|
|
|
}
|
2015-08-21 05:33:25 +00:00
|
|
|
|
|
|
|
TableReader* table_reader = nullptr;
|
2014-04-17 21:07:05 +00:00
|
|
|
Cache::Handle* handle = nullptr;
|
2015-08-21 05:33:25 +00:00
|
|
|
bool create_new_table_reader =
|
|
|
|
(for_compaction && ioptions_.new_table_reader_for_compaction_inputs);
|
|
|
|
if (create_new_table_reader) {
|
|
|
|
unique_ptr<TableReader> table_reader_unique_ptr;
|
2015-08-26 22:25:59 +00:00
|
|
|
Status s = GetTableReader(
|
|
|
|
env_options, icomparator, fd, /* sequential mode */ true,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
/* record stats */ false, nullptr, &table_reader_unique_ptr,
|
|
|
|
false /* skip_filters */, level);
|
2014-04-17 22:14:04 +00:00
|
|
|
if (!s.ok()) {
|
2015-10-12 22:06:38 +00:00
|
|
|
return NewErrorInternalIterator(s, arena);
|
2014-04-17 22:14:04 +00:00
|
|
|
}
|
2015-08-21 05:33:25 +00:00
|
|
|
table_reader = table_reader_unique_ptr.release();
|
|
|
|
} else {
|
|
|
|
table_reader = fd.table_reader;
|
|
|
|
if (table_reader == nullptr) {
|
Skip bottom-level filter block caching when hit-optimized
Summary:
When Get() or NewIterator() trigger file loads, skip caching the filter block if
(1) optimize_filters_for_hits is set and (2) the file is on the bottommost
level. Also skip checking filters under the same conditions, which means that
for a preloaded file or a file that was trivially-moved to the bottom level, its
filter block will eventually expire from the cache.
- added parameters/instance variables in various places in order to propagate the config ("skip_filters") from version_set to block_based_table_reader
- in BlockBasedTable::Rep, this optimization prevents filter from being loaded when the file is opened simply by setting filter_policy = nullptr
- in BlockBasedTable::Get/BlockBasedTable::NewIterator, this optimization prevents filter from being used (even if it was loaded already) by setting filter = nullptr
Test Plan:
updated unit test:
$ ./db_test --gtest_filter=DBTest.OptimizeFiltersForHits
will also run 'make check'
Reviewers: sdong, igor, paultuckfield, anthony, rven, kradhakrishnan, IslamAbdelRahman, yhchiang
Reviewed By: yhchiang
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D51633
2015-12-23 18:15:07 +00:00
|
|
|
Status s = FindTable(env_options, icomparator, fd, &handle,
|
|
|
|
options.read_tier == kBlockCacheTier /* no_io */,
|
|
|
|
!for_compaction /* record read_stats */,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
file_read_hist, skip_filters, level);
|
2015-08-21 05:33:25 +00:00
|
|
|
if (!s.ok()) {
|
2015-10-12 22:06:38 +00:00
|
|
|
return NewErrorInternalIterator(s, arena);
|
2015-08-21 05:33:25 +00:00
|
|
|
}
|
|
|
|
table_reader = GetTableReaderFromHandle(handle);
|
|
|
|
}
|
2014-01-07 04:29:17 +00:00
|
|
|
}
|
2011-03-18 22:37:00 +00:00
|
|
|
|
Skip bottom-level filter block caching when hit-optimized
Summary:
When Get() or NewIterator() trigger file loads, skip caching the filter block if
(1) optimize_filters_for_hits is set and (2) the file is on the bottommost
level. Also skip checking filters under the same conditions, which means that
for a preloaded file or a file that was trivially-moved to the bottom level, its
filter block will eventually expire from the cache.
- added parameters/instance variables in various places in order to propagate the config ("skip_filters") from version_set to block_based_table_reader
- in BlockBasedTable::Rep, this optimization prevents filter from being loaded when the file is opened simply by setting filter_policy = nullptr
- in BlockBasedTable::Get/BlockBasedTable::NewIterator, this optimization prevents filter from being used (even if it was loaded already) by setting filter = nullptr
Test Plan:
updated unit test:
$ ./db_test --gtest_filter=DBTest.OptimizeFiltersForHits
will also run 'make check'
Reviewers: sdong, igor, paultuckfield, anthony, rven, kradhakrishnan, IslamAbdelRahman, yhchiang
Reviewed By: yhchiang
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D51633
2015-12-23 18:15:07 +00:00
|
|
|
InternalIterator* result =
|
|
|
|
table_reader->NewIterator(options, arena, skip_filters);
|
2015-08-21 05:33:25 +00:00
|
|
|
|
|
|
|
if (create_new_table_reader) {
|
|
|
|
assert(handle == nullptr);
|
|
|
|
result->RegisterCleanup(&DeleteTableReader, table_reader, nullptr);
|
|
|
|
} else if (handle != nullptr) {
|
2014-02-06 23:42:16 +00:00
|
|
|
result->RegisterCleanup(&UnrefEntry, cache_, handle);
|
2014-01-07 04:29:17 +00:00
|
|
|
}
|
2013-05-17 22:53:01 +00:00
|
|
|
|
|
|
|
if (for_compaction) {
|
2013-10-30 17:52:33 +00:00
|
|
|
table_reader->SetupForCompaction();
|
2013-05-17 22:53:01 +00:00
|
|
|
}
|
2015-08-21 05:33:25 +00:00
|
|
|
if (table_reader_ptr != nullptr) {
|
|
|
|
*table_reader_ptr = table_reader;
|
|
|
|
}
|
2013-05-17 22:53:01 +00:00
|
|
|
|
2011-03-18 22:37:00 +00:00
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2012-04-17 15:36:46 +00:00
|
|
|
Status TableCache::Get(const ReadOptions& options,
|
2014-01-27 21:53:22 +00:00
|
|
|
const InternalKeyComparator& internal_comparator,
|
2014-09-29 18:09:09 +00:00
|
|
|
const FileDescriptor& fd, const Slice& k,
|
Skip bottom-level filter block caching when hit-optimized
Summary:
When Get() or NewIterator() trigger file loads, skip caching the filter block if
(1) optimize_filters_for_hits is set and (2) the file is on the bottommost
level. Also skip checking filters under the same conditions, which means that
for a preloaded file or a file that was trivially-moved to the bottom level, its
filter block will eventually expire from the cache.
- added parameters/instance variables in various places in order to propagate the config ("skip_filters") from version_set to block_based_table_reader
- in BlockBasedTable::Rep, this optimization prevents filter from being loaded when the file is opened simply by setting filter_policy = nullptr
- in BlockBasedTable::Get/BlockBasedTable::NewIterator, this optimization prevents filter from being used (even if it was loaded already) by setting filter = nullptr
Test Plan:
updated unit test:
$ ./db_test --gtest_filter=DBTest.OptimizeFiltersForHits
will also run 'make check'
Reviewers: sdong, igor, paultuckfield, anthony, rven, kradhakrishnan, IslamAbdelRahman, yhchiang
Reviewed By: yhchiang
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D51633
2015-12-23 18:15:07 +00:00
|
|
|
GetContext* get_context, HistogramImpl* file_read_hist,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
bool skip_filters, int level) {
|
2014-06-13 22:54:19 +00:00
|
|
|
TableReader* t = fd.table_reader;
|
2014-01-07 04:29:17 +00:00
|
|
|
Status s;
|
2014-04-17 21:07:05 +00:00
|
|
|
Cache::Handle* handle = nullptr;
|
2015-06-23 17:25:45 +00:00
|
|
|
std::string* row_cache_entry = nullptr;
|
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
IterKey row_cache_key;
|
|
|
|
std::string row_cache_entry_buffer;
|
|
|
|
|
Use SST files for Transaction conflict detection
Summary:
Currently, transactions can fail even if there is no actual write conflict. This is due to relying on only the memtables to check for write-conflicts. Users have to tune memtable settings to try to avoid this, but it's hard to figure out exactly how to tune these settings.
With this diff, TransactionDB will use both memtables and SST files to determine if there are any write conflicts. This relies on the fact that BlockBasedTable stores sequence numbers for all writes that happen after any open snapshot. Also, D50295 is needed to prevent SingleDelete from disappearing writes (the TODOs in this test code will be fixed once the other diff is approved and merged).
Note that Optimistic transactions will still rely on tuning memtable settings as we do not want to read from SST while on the write thread. Also, memtable settings can still be used to reduce how often TransactionDB needs to read SST files.
Test Plan: unit tests, db bench
Reviewers: rven, yhchiang, kradhakrishnan, IslamAbdelRahman, sdong
Reviewed By: sdong
Subscribers: dhruba, leveldb, yoshinorim
Differential Revision: https://reviews.facebook.net/D50475
2015-10-15 23:37:15 +00:00
|
|
|
// Check row cache if enabled. Since row cache does not currently store
|
|
|
|
// sequence numbers, we cannot use it if we need to fetch the sequence.
|
|
|
|
if (ioptions_.row_cache && !get_context->NeedToReadSequence()) {
|
2015-06-23 17:25:45 +00:00
|
|
|
uint64_t fd_number = fd.GetNumber();
|
|
|
|
auto user_key = ExtractUserKey(k);
|
|
|
|
// We use the user key as cache key instead of the internal key,
|
|
|
|
// otherwise the whole cache would be invalidated every time the
|
|
|
|
// sequence key increases. However, to support caching snapshot
|
|
|
|
// reads, we append the sequence number (incremented by 1 to
|
|
|
|
// distinguish from 0) only in this case.
|
|
|
|
uint64_t seq_no =
|
|
|
|
options.snapshot == nullptr ? 0 : 1 + GetInternalKeySeqno(k);
|
|
|
|
|
|
|
|
// Compute row cache key.
|
|
|
|
row_cache_key.TrimAppend(row_cache_key.Size(), row_cache_id_.data(),
|
|
|
|
row_cache_id_.size());
|
|
|
|
AppendVarint64(&row_cache_key, fd_number);
|
|
|
|
AppendVarint64(&row_cache_key, seq_no);
|
|
|
|
row_cache_key.TrimAppend(row_cache_key.Size(), user_key.data(),
|
|
|
|
user_key.size());
|
|
|
|
|
|
|
|
if (auto row_handle = ioptions_.row_cache->Lookup(row_cache_key.GetKey())) {
|
|
|
|
auto found_row_cache_entry = static_cast<const std::string*>(
|
|
|
|
ioptions_.row_cache->Value(row_handle));
|
|
|
|
replayGetContextLog(*found_row_cache_entry, user_key, get_context);
|
|
|
|
ioptions_.row_cache->Release(row_handle);
|
|
|
|
RecordTick(ioptions_.statistics, ROW_CACHE_HIT);
|
|
|
|
return Status::OK();
|
|
|
|
}
|
|
|
|
|
|
|
|
// Not found, setting up the replay log.
|
|
|
|
RecordTick(ioptions_.statistics, ROW_CACHE_MISS);
|
|
|
|
row_cache_entry = &row_cache_entry_buffer;
|
|
|
|
}
|
|
|
|
#endif // ROCKSDB_LITE
|
|
|
|
|
2014-04-17 21:07:05 +00:00
|
|
|
if (!t) {
|
2014-09-04 23:18:36 +00:00
|
|
|
s = FindTable(env_options_, internal_comparator, fd, &handle,
|
2015-09-02 17:25:20 +00:00
|
|
|
options.read_tier == kBlockCacheTier /* no_io */,
|
Adding pin_l0_filter_and_index_blocks_in_cache feature.
Summary:
When a block based table file is opened, if prefetch_index_and_filter is true, it will prefetch the index and filter blocks, putting them into the block cache.
What this feature adds: when a L0 block based table file is opened, if pin_l0_filter_and_index_blocks_in_cache is true in the options (and prefetch_index_and_filter is true), then the filter and index blocks aren't released back to the block cache at the end of BlockBasedTableReader::Open(). Instead the table reader takes ownership of them, hence pinning them, ie. the LRU cache will never push them out. Meanwhile in the table reader, further accesses will not hit the block cache, thus avoiding lock contention.
When the table reader is destroyed, it releases the pinned blocks (if there were any). This has to happen before the cache is destroyed, so I had to introduce a TableReader::Close(), to guarantee the order of destruction.
Test Plan:
Added two unit tests for this. Existing unit tests run fine (default is pin_l0_filter_and_index_blocks_in_cache=false).
DISABLE_JEMALLOC=1 OPT=-g make all valgrind_check -j32
Mac: OK.
Linux: with D55287 patched in it's OK.
Reviewers: sdong
Reviewed By: sdong
Subscribers: andrewkr, leveldb, dhruba
Differential Revision: https://reviews.facebook.net/D54801
2016-03-17 22:40:01 +00:00
|
|
|
true /* record_read_stats */, file_read_hist, skip_filters,
|
|
|
|
level);
|
2014-04-17 22:14:04 +00:00
|
|
|
if (s.ok()) {
|
|
|
|
t = GetTableReaderFromHandle(handle);
|
|
|
|
}
|
2014-01-07 04:29:17 +00:00
|
|
|
}
|
2012-04-17 15:36:46 +00:00
|
|
|
if (s.ok()) {
|
2015-06-23 17:25:45 +00:00
|
|
|
get_context->SetReplayLog(row_cache_entry); // nullptr if no cache.
|
Skip bottom-level filter block caching when hit-optimized
Summary:
When Get() or NewIterator() trigger file loads, skip caching the filter block if
(1) optimize_filters_for_hits is set and (2) the file is on the bottommost
level. Also skip checking filters under the same conditions, which means that
for a preloaded file or a file that was trivially-moved to the bottom level, its
filter block will eventually expire from the cache.
- added parameters/instance variables in various places in order to propagate the config ("skip_filters") from version_set to block_based_table_reader
- in BlockBasedTable::Rep, this optimization prevents filter from being loaded when the file is opened simply by setting filter_policy = nullptr
- in BlockBasedTable::Get/BlockBasedTable::NewIterator, this optimization prevents filter from being used (even if it was loaded already) by setting filter = nullptr
Test Plan:
updated unit test:
$ ./db_test --gtest_filter=DBTest.OptimizeFiltersForHits
will also run 'make check'
Reviewers: sdong, igor, paultuckfield, anthony, rven, kradhakrishnan, IslamAbdelRahman, yhchiang
Reviewed By: yhchiang
Subscribers: leveldb
Differential Revision: https://reviews.facebook.net/D51633
2015-12-23 18:15:07 +00:00
|
|
|
s = t->Get(options, k, get_context, skip_filters);
|
2015-06-23 17:25:45 +00:00
|
|
|
get_context->SetReplayLog(nullptr);
|
2014-04-17 21:07:05 +00:00
|
|
|
if (handle != nullptr) {
|
2014-01-07 04:29:17 +00:00
|
|
|
ReleaseHandle(handle);
|
|
|
|
}
|
2016-02-09 19:20:22 +00:00
|
|
|
} else if (options.read_tier == kBlockCacheTier && s.IsIncomplete()) {
|
2015-06-23 17:25:45 +00:00
|
|
|
// Couldn't find Table in cache but treat as kFound if no_io set
|
2014-09-29 18:09:09 +00:00
|
|
|
get_context->MarkKeyMayExist();
|
2013-07-12 23:56:52 +00:00
|
|
|
return Status::OK();
|
2012-04-17 15:36:46 +00:00
|
|
|
}
|
2015-06-23 17:25:45 +00:00
|
|
|
|
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
// Put the replay log in row cache only if something was found.
|
|
|
|
if (s.ok() && row_cache_entry && !row_cache_entry->empty()) {
|
|
|
|
size_t charge =
|
|
|
|
row_cache_key.Size() + row_cache_entry->size() + sizeof(std::string);
|
|
|
|
void* row_ptr = new std::string(std::move(*row_cache_entry));
|
2016-03-11 01:35:19 +00:00
|
|
|
ioptions_.row_cache->Insert(row_cache_key.GetKey(), row_ptr, charge,
|
|
|
|
&DeleteEntry<std::string>);
|
2015-06-23 17:25:45 +00:00
|
|
|
}
|
|
|
|
#endif // ROCKSDB_LITE
|
|
|
|
|
2012-04-17 15:36:46 +00:00
|
|
|
return s;
|
|
|
|
}
|
2014-09-04 23:18:36 +00:00
|
|
|
|
2014-02-14 00:28:21 +00:00
|
|
|
Status TableCache::GetTableProperties(
|
2014-09-04 23:18:36 +00:00
|
|
|
const EnvOptions& env_options,
|
2014-06-13 22:54:19 +00:00
|
|
|
const InternalKeyComparator& internal_comparator, const FileDescriptor& fd,
|
2014-02-14 00:28:21 +00:00
|
|
|
std::shared_ptr<const TableProperties>* properties, bool no_io) {
|
|
|
|
Status s;
|
2014-06-13 22:54:19 +00:00
|
|
|
auto table_reader = fd.table_reader;
|
2014-02-14 00:28:21 +00:00
|
|
|
// table already been pre-loaded?
|
2014-04-17 21:07:05 +00:00
|
|
|
if (table_reader) {
|
|
|
|
*properties = table_reader->GetTableProperties();
|
|
|
|
|
2014-02-14 00:28:21 +00:00
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
2014-04-17 21:07:05 +00:00
|
|
|
Cache::Handle* table_handle = nullptr;
|
2014-09-04 23:18:36 +00:00
|
|
|
s = FindTable(env_options, internal_comparator, fd, &table_handle, no_io);
|
2014-02-14 00:28:21 +00:00
|
|
|
if (!s.ok()) {
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
assert(table_handle);
|
|
|
|
auto table = GetTableReaderFromHandle(table_handle);
|
|
|
|
*properties = table->GetTableProperties();
|
|
|
|
ReleaseHandle(table_handle);
|
|
|
|
return s;
|
|
|
|
}
|
2012-04-17 15:36:46 +00:00
|
|
|
|
2014-08-05 18:27:34 +00:00
|
|
|
size_t TableCache::GetMemoryUsageByTableReader(
|
2014-09-04 23:18:36 +00:00
|
|
|
const EnvOptions& env_options,
|
2014-08-05 18:27:34 +00:00
|
|
|
const InternalKeyComparator& internal_comparator,
|
|
|
|
const FileDescriptor& fd) {
|
|
|
|
Status s;
|
|
|
|
auto table_reader = fd.table_reader;
|
|
|
|
// table already been pre-loaded?
|
|
|
|
if (table_reader) {
|
|
|
|
return table_reader->ApproximateMemoryUsage();
|
|
|
|
}
|
|
|
|
|
|
|
|
Cache::Handle* table_handle = nullptr;
|
2014-09-04 23:18:36 +00:00
|
|
|
s = FindTable(env_options, internal_comparator, fd, &table_handle, true);
|
2014-08-05 18:27:34 +00:00
|
|
|
if (!s.ok()) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
assert(table_handle);
|
|
|
|
auto table = GetTableReaderFromHandle(table_handle);
|
|
|
|
auto ret = table->ApproximateMemoryUsage();
|
|
|
|
ReleaseHandle(table_handle);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
[CF] Rethink table cache
Summary:
Adapting table cache to column families is interesting. We want table cache to be global LRU, so if some column families are use not as often as others, we want them to be evicted from cache. However, current TableCache object also constructs tables on its own. If table is not found in the cache, TableCache automatically creates new table. We want each column family to be able to specify different table factory.
To solve the problem, we still have a single LRU, but we provide the LRUCache object to TableCache on construction. We have one TableCache per column family, but the underyling cache is shared by all TableCache objects.
This allows us to have a global LRU, but still be able to support different table factories for different column families. Also, in the future it will also be able to support different directories for different column families.
Test Plan: make check
Reviewers: dhruba, haobo, kailiu, sdong
CC: leveldb
Differential Revision: https://reviews.facebook.net/D15915
2014-02-05 17:07:55 +00:00
|
|
|
void TableCache::Evict(Cache* cache, uint64_t file_number) {
|
|
|
|
cache->Erase(GetSliceForFileNumber(&file_number));
|
2011-03-18 22:37:00 +00:00
|
|
|
}
|
|
|
|
|
2013-10-04 04:49:15 +00:00
|
|
|
} // namespace rocksdb
|