mirror of
https://github.com/facebook/rocksdb.git
synced 2024-11-25 14:31:35 +00:00
9d70cce047
Summary: Adding option to save PlainTable index and bloom filter in SST file. If there is no bloom block and/or index block, PlainTableReader builds new ones. Otherwise PlainTableReader just use these blocks. Test Plan: make all check Reviewers: sdong Reviewed By: sdong Subscribers: leveldb Differential Revision: https://reviews.facebook.net/D19527
154 lines
5 KiB
C++
154 lines
5 KiB
C++
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under the BSD-style license found in the
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
|
|
#pragma once
|
|
|
|
#include <string>
|
|
|
|
#include "rocksdb/slice.h"
|
|
|
|
#include <util/arena.h>
|
|
#include <port/port_posix.h>
|
|
|
|
#include <atomic>
|
|
#include <memory>
|
|
|
|
namespace rocksdb {
|
|
|
|
class Slice;
|
|
class Logger;
|
|
|
|
class DynamicBloom {
|
|
public:
|
|
// arena: pass arena to bloom filter, hence trace the usage of memory
|
|
// total_bits: fixed total bits for the bloom
|
|
// num_probes: number of hash probes for a single key
|
|
// locality: If positive, optimize for cache line locality, 0 otherwise.
|
|
// hash_func: customized hash function
|
|
// huge_page_tlb_size: if >0, try to allocate bloom bytes from huge page TLB
|
|
// withi this page size. Need to reserve huge pages for
|
|
// it to be allocated, like:
|
|
// sysctl -w vm.nr_hugepages=20
|
|
// See linux doc Documentation/vm/hugetlbpage.txt
|
|
explicit DynamicBloom(Arena* arena,
|
|
uint32_t total_bits, uint32_t locality = 0,
|
|
uint32_t num_probes = 6,
|
|
uint32_t (*hash_func)(const Slice& key) = nullptr,
|
|
size_t huge_page_tlb_size = 0,
|
|
Logger* logger = nullptr);
|
|
|
|
explicit DynamicBloom(uint32_t num_probes = 6,
|
|
uint32_t (*hash_func)(const Slice& key) = nullptr);
|
|
|
|
void SetTotalBits(Arena* arena, uint32_t total_bits, uint32_t locality,
|
|
size_t huge_page_tlb_size, Logger* logger);
|
|
|
|
~DynamicBloom() {}
|
|
|
|
// Assuming single threaded access to this function.
|
|
void Add(const Slice& key);
|
|
|
|
// Assuming single threaded access to this function.
|
|
void AddHash(uint32_t hash);
|
|
|
|
// Multithreaded access to this function is OK
|
|
bool MayContain(const Slice& key) const;
|
|
|
|
// Multithreaded access to this function is OK
|
|
bool MayContainHash(uint32_t hash) const;
|
|
|
|
void Prefetch(uint32_t h);
|
|
|
|
uint32_t GetNumBlocks() const { return kNumBlocks; }
|
|
|
|
Slice GetRawData() const {
|
|
return Slice(reinterpret_cast<char*>(data_), GetTotalBits() / 8);
|
|
}
|
|
|
|
void SetRawData(unsigned char* raw_data, uint32_t total_bits,
|
|
uint32_t num_blocks = 0);
|
|
|
|
uint32_t GetTotalBits() const { return kTotalBits; }
|
|
|
|
bool IsInitialized() const { return kNumBlocks > 0 || kTotalBits > 0; }
|
|
|
|
private:
|
|
uint32_t kTotalBits;
|
|
uint32_t kNumBlocks;
|
|
const uint32_t kNumProbes;
|
|
|
|
uint32_t (*hash_func_)(const Slice& key);
|
|
unsigned char* data_;
|
|
unsigned char* raw_;
|
|
};
|
|
|
|
inline void DynamicBloom::Add(const Slice& key) { AddHash(hash_func_(key)); }
|
|
|
|
inline bool DynamicBloom::MayContain(const Slice& key) const {
|
|
return (MayContainHash(hash_func_(key)));
|
|
}
|
|
|
|
inline void DynamicBloom::Prefetch(uint32_t h) {
|
|
if (kNumBlocks != 0) {
|
|
uint32_t b = ((h >> 11 | (h << 21)) % kNumBlocks) * (CACHE_LINE_SIZE * 8);
|
|
PREFETCH(&(data_[b]), 0, 3);
|
|
}
|
|
}
|
|
|
|
inline bool DynamicBloom::MayContainHash(uint32_t h) const {
|
|
assert(IsInitialized());
|
|
const uint32_t delta = (h >> 17) | (h << 15); // Rotate right 17 bits
|
|
if (kNumBlocks != 0) {
|
|
uint32_t b = ((h >> 11 | (h << 21)) % kNumBlocks) * (CACHE_LINE_SIZE * 8);
|
|
for (uint32_t i = 0; i < kNumProbes; ++i) {
|
|
// Since CACHE_LINE_SIZE is defined as 2^n, this line will be optimized
|
|
// to a simple and operation by compiler.
|
|
const uint32_t bitpos = b + (h % (CACHE_LINE_SIZE * 8));
|
|
if (((data_[bitpos / 8]) & (1 << (bitpos % 8))) == 0) {
|
|
return false;
|
|
}
|
|
// Rotate h so that we don't reuse the same bytes.
|
|
h = h / (CACHE_LINE_SIZE * 8) +
|
|
(h % (CACHE_LINE_SIZE * 8)) * (0x20000000U / CACHE_LINE_SIZE);
|
|
h += delta;
|
|
}
|
|
} else {
|
|
for (uint32_t i = 0; i < kNumProbes; ++i) {
|
|
const uint32_t bitpos = h % kTotalBits;
|
|
if (((data_[bitpos / 8]) & (1 << (bitpos % 8))) == 0) {
|
|
return false;
|
|
}
|
|
h += delta;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
inline void DynamicBloom::AddHash(uint32_t h) {
|
|
assert(IsInitialized());
|
|
const uint32_t delta = (h >> 17) | (h << 15); // Rotate right 17 bits
|
|
if (kNumBlocks != 0) {
|
|
uint32_t b = ((h >> 11 | (h << 21)) % kNumBlocks) * (CACHE_LINE_SIZE * 8);
|
|
for (uint32_t i = 0; i < kNumProbes; ++i) {
|
|
// Since CACHE_LINE_SIZE is defined as 2^n, this line will be optimized
|
|
// to a simple and operation by compiler.
|
|
const uint32_t bitpos = b + (h % (CACHE_LINE_SIZE * 8));
|
|
data_[bitpos / 8] |= (1 << (bitpos % 8));
|
|
// Rotate h so that we don't reuse the same bytes.
|
|
h = h / (CACHE_LINE_SIZE * 8) +
|
|
(h % (CACHE_LINE_SIZE * 8)) * (0x20000000U / CACHE_LINE_SIZE);
|
|
h += delta;
|
|
}
|
|
} else {
|
|
for (uint32_t i = 0; i < kNumProbes; ++i) {
|
|
const uint32_t bitpos = h % kTotalBits;
|
|
data_[bitpos / 8] |= (1 << (bitpos % 8));
|
|
h += delta;
|
|
}
|
|
}
|
|
}
|
|
|
|
} // rocksdb
|