2016-04-22 23:18:33 +00:00
|
|
|
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
2017-07-15 23:03:42 +00:00
|
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
|
|
// (found in the LICENSE.Apache file in the root directory).
|
2016-04-22 23:18:33 +00:00
|
|
|
#ifndef ROCKSDB_LITE
|
|
|
|
|
|
|
|
#include "utilities/persistent_cache/block_cache_tier_file.h"
|
|
|
|
|
2016-08-03 00:15:18 +00:00
|
|
|
#ifndef OS_WIN
|
2016-04-22 23:18:33 +00:00
|
|
|
#include <unistd.h>
|
2016-08-03 00:15:18 +00:00
|
|
|
#endif
|
2016-12-16 19:17:26 +00:00
|
|
|
#include <functional>
|
2016-04-22 23:18:33 +00:00
|
|
|
#include <memory>
|
|
|
|
#include <vector>
|
|
|
|
|
Introduce a new storage specific Env API (#5761)
Summary:
The current Env API encompasses both storage/file operations, as well as OS related operations. Most of the APIs return a Status, which does not have enough metadata about an error, such as whether its retry-able or not, scope (i.e fault domain) of the error etc., that may be required in order to properly handle a storage error. The file APIs also do not provide enough control over the IO SLA, such as timeout, prioritization, hinting about placement and redundancy etc.
This PR separates out the file/storage APIs from Env into a new FileSystem class. The APIs are updated to return an IOStatus with metadata about the error, as well as to take an IOOptions structure as input in order to allow more control over the IO.
The user can set both ```options.env``` and ```options.file_system``` to specify that RocksDB should use the former for OS related operations and the latter for storage operations. Internally, a ```CompositeEnvWrapper``` has been introduced that inherits from ```Env``` and redirects individual methods to either an ```Env``` implementation or the ```FileSystem``` as appropriate. When options are sanitized during ```DB::Open```, ```options.env``` is replaced with a newly allocated ```CompositeEnvWrapper``` instance if both env and file_system have been specified. This way, the rest of the RocksDB code can continue to function as before.
This PR also ports PosixEnv to the new API by splitting it into two - PosixEnv and PosixFileSystem. PosixEnv is defined as a sub-class of CompositeEnvWrapper, and threading/time functions are overridden with Posix specific implementations in order to avoid an extra level of indirection.
The ```CompositeEnvWrapper``` translates ```IOStatus``` return code to ```Status```, and sets the severity to ```kSoftError``` if the io_status is retryable. The error handling code in RocksDB can then recover the DB automatically.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5761
Differential Revision: D18868376
Pulled By: anand1976
fbshipit-source-id: 39efe18a162ea746fabac6360ff529baba48486f
2019-12-13 22:47:08 +00:00
|
|
|
#include "env/composite_env_wrapper.h"
|
2019-06-01 00:19:43 +00:00
|
|
|
#include "logging/logging.h"
|
2017-02-06 22:43:55 +00:00
|
|
|
#include "port/port.h"
|
2017-03-16 02:22:52 +00:00
|
|
|
#include "util/crc32c.h"
|
2016-04-22 23:18:33 +00:00
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
namespace ROCKSDB_NAMESPACE {
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
//
|
|
|
|
// File creation factories
|
|
|
|
//
|
|
|
|
Status NewWritableCacheFile(Env* const env, const std::string& filepath,
|
|
|
|
std::unique_ptr<WritableFile>* file,
|
|
|
|
const bool use_direct_writes = false) {
|
|
|
|
EnvOptions opt;
|
|
|
|
opt.use_direct_writes = use_direct_writes;
|
|
|
|
Status s = env->NewWritableFile(filepath, file, opt);
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
Status NewRandomAccessCacheFile(Env* const env, const std::string& filepath,
|
|
|
|
std::unique_ptr<RandomAccessFile>* file,
|
|
|
|
const bool use_direct_reads = true) {
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(env);
|
|
|
|
|
2016-04-22 23:18:33 +00:00
|
|
|
EnvOptions opt;
|
|
|
|
opt.use_direct_reads = use_direct_reads;
|
|
|
|
Status s = env->NewRandomAccessFile(filepath, file, opt);
|
|
|
|
return s;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// BlockCacheFile
|
|
|
|
//
|
2016-07-19 23:36:25 +00:00
|
|
|
Status BlockCacheFile::Delete(uint64_t* size) {
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(env_);
|
|
|
|
|
2016-04-22 23:18:33 +00:00
|
|
|
Status status = env_->GetFileSize(Path(), size);
|
|
|
|
if (!status.ok()) {
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
return env_->DeleteFile(Path());
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// CacheRecord
|
|
|
|
//
|
|
|
|
// Cache record represents the record on disk
|
|
|
|
//
|
|
|
|
// +--------+---------+----------+------------+---------------+-------------+
|
|
|
|
// | magic | crc | key size | value size | key data | value data |
|
|
|
|
// +--------+---------+----------+------------+---------------+-------------+
|
|
|
|
// <-- 4 --><-- 4 --><-- 4 --><-- 4 --><-- key size --><-- v-size -->
|
|
|
|
//
|
|
|
|
struct CacheRecordHeader {
|
2017-11-28 21:15:20 +00:00
|
|
|
CacheRecordHeader()
|
|
|
|
: magic_(0), crc_(0), key_size_(0), val_size_(0) {}
|
2016-04-22 23:18:33 +00:00
|
|
|
CacheRecordHeader(const uint32_t magic, const uint32_t key_size,
|
|
|
|
const uint32_t val_size)
|
|
|
|
: magic_(magic), crc_(0), key_size_(key_size), val_size_(val_size) {}
|
|
|
|
|
|
|
|
uint32_t magic_;
|
|
|
|
uint32_t crc_;
|
|
|
|
uint32_t key_size_;
|
|
|
|
uint32_t val_size_;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct CacheRecord {
|
|
|
|
CacheRecord() {}
|
|
|
|
CacheRecord(const Slice& key, const Slice& val)
|
|
|
|
: hdr_(MAGIC, static_cast<uint32_t>(key.size()),
|
|
|
|
static_cast<uint32_t>(val.size())),
|
|
|
|
key_(key),
|
|
|
|
val_(val) {
|
|
|
|
hdr_.crc_ = ComputeCRC();
|
|
|
|
}
|
|
|
|
|
|
|
|
uint32_t ComputeCRC() const;
|
|
|
|
bool Serialize(std::vector<CacheWriteBuffer*>* bufs, size_t* woff);
|
|
|
|
bool Deserialize(const Slice& buf);
|
|
|
|
|
|
|
|
static uint32_t CalcSize(const Slice& key, const Slice& val) {
|
|
|
|
return static_cast<uint32_t>(sizeof(CacheRecordHeader) + key.size() +
|
|
|
|
val.size());
|
|
|
|
}
|
|
|
|
|
|
|
|
static const uint32_t MAGIC = 0xfefa;
|
|
|
|
|
|
|
|
bool Append(std::vector<CacheWriteBuffer*>* bufs, size_t* woff,
|
|
|
|
const char* data, const size_t size);
|
|
|
|
|
|
|
|
CacheRecordHeader hdr_;
|
|
|
|
Slice key_;
|
|
|
|
Slice val_;
|
|
|
|
};
|
|
|
|
|
|
|
|
static_assert(sizeof(CacheRecordHeader) == 16, "DataHeader is not aligned");
|
|
|
|
|
|
|
|
uint32_t CacheRecord::ComputeCRC() const {
|
|
|
|
uint32_t crc = 0;
|
|
|
|
CacheRecordHeader tmp = hdr_;
|
|
|
|
tmp.crc_ = 0;
|
|
|
|
crc = crc32c::Extend(crc, reinterpret_cast<const char*>(&tmp), sizeof(tmp));
|
|
|
|
crc = crc32c::Extend(crc, reinterpret_cast<const char*>(key_.data()),
|
|
|
|
key_.size());
|
|
|
|
crc = crc32c::Extend(crc, reinterpret_cast<const char*>(val_.data()),
|
|
|
|
val_.size());
|
|
|
|
return crc;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool CacheRecord::Serialize(std::vector<CacheWriteBuffer*>* bufs,
|
|
|
|
size_t* woff) {
|
|
|
|
assert(bufs->size());
|
|
|
|
return Append(bufs, woff, reinterpret_cast<const char*>(&hdr_),
|
|
|
|
sizeof(hdr_)) &&
|
|
|
|
Append(bufs, woff, reinterpret_cast<const char*>(key_.data()),
|
|
|
|
key_.size()) &&
|
|
|
|
Append(bufs, woff, reinterpret_cast<const char*>(val_.data()),
|
|
|
|
val_.size());
|
|
|
|
}
|
|
|
|
|
|
|
|
bool CacheRecord::Append(std::vector<CacheWriteBuffer*>* bufs, size_t* woff,
|
|
|
|
const char* data, const size_t data_size) {
|
|
|
|
assert(*woff < bufs->size());
|
|
|
|
|
|
|
|
const char* p = data;
|
|
|
|
size_t size = data_size;
|
|
|
|
|
|
|
|
while (size && *woff < bufs->size()) {
|
|
|
|
CacheWriteBuffer* buf = (*bufs)[*woff];
|
|
|
|
const size_t free = buf->Free();
|
|
|
|
if (size <= free) {
|
|
|
|
buf->Append(p, size);
|
|
|
|
size = 0;
|
|
|
|
} else {
|
|
|
|
buf->Append(p, free);
|
|
|
|
p += free;
|
|
|
|
size -= free;
|
|
|
|
assert(!buf->Free());
|
|
|
|
assert(buf->Used() == buf->Capacity());
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!buf->Free()) {
|
|
|
|
*woff += 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(!size);
|
|
|
|
|
|
|
|
return !size;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool CacheRecord::Deserialize(const Slice& data) {
|
|
|
|
assert(data.size() >= sizeof(CacheRecordHeader));
|
|
|
|
if (data.size() < sizeof(CacheRecordHeader)) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
memcpy(&hdr_, data.data(), sizeof(hdr_));
|
|
|
|
|
|
|
|
assert(hdr_.key_size_ + hdr_.val_size_ + sizeof(hdr_) == data.size());
|
|
|
|
if (hdr_.key_size_ + hdr_.val_size_ + sizeof(hdr_) != data.size()) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
key_ = Slice(data.data_ + sizeof(hdr_), hdr_.key_size_);
|
|
|
|
val_ = Slice(key_.data_ + hdr_.key_size_, hdr_.val_size_);
|
|
|
|
|
|
|
|
if (!(hdr_.magic_ == MAGIC && ComputeCRC() == hdr_.crc_)) {
|
|
|
|
fprintf(stderr, "** magic %d ** \n", hdr_.magic_);
|
|
|
|
fprintf(stderr, "** key_size %d ** \n", hdr_.key_size_);
|
|
|
|
fprintf(stderr, "** val_size %d ** \n", hdr_.val_size_);
|
|
|
|
fprintf(stderr, "** key %s ** \n", key_.ToString().c_str());
|
|
|
|
fprintf(stderr, "** val %s ** \n", val_.ToString().c_str());
|
|
|
|
for (size_t i = 0; i < hdr_.val_size_; ++i) {
|
|
|
|
fprintf(stderr, "%d.", (uint8_t)val_.data()[i]);
|
|
|
|
}
|
|
|
|
fprintf(stderr, "\n** cksum %d != %d **", hdr_.crc_, ComputeCRC());
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(hdr_.magic_ == MAGIC && ComputeCRC() == hdr_.crc_);
|
|
|
|
return hdr_.magic_ == MAGIC && ComputeCRC() == hdr_.crc_;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// RandomAccessFile
|
|
|
|
//
|
|
|
|
|
2016-05-03 18:05:42 +00:00
|
|
|
bool RandomAccessCacheFile::Open(const bool enable_direct_reads) {
|
2016-04-22 23:18:33 +00:00
|
|
|
WriteLock _(&rwlock_);
|
2016-05-03 18:05:42 +00:00
|
|
|
return OpenImpl(enable_direct_reads);
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
|
|
|
|
2016-05-03 18:05:42 +00:00
|
|
|
bool RandomAccessCacheFile::OpenImpl(const bool enable_direct_reads) {
|
2016-04-22 23:18:33 +00:00
|
|
|
rwlock_.AssertHeld();
|
|
|
|
|
2017-03-16 02:22:52 +00:00
|
|
|
ROCKS_LOG_DEBUG(log_, "Opening cache file %s", Path().c_str());
|
2016-04-22 23:18:33 +00:00
|
|
|
|
2017-02-27 20:59:34 +00:00
|
|
|
std::unique_ptr<RandomAccessFile> file;
|
|
|
|
Status status =
|
|
|
|
NewRandomAccessCacheFile(env_, Path(), &file, enable_direct_reads);
|
2016-04-22 23:18:33 +00:00
|
|
|
if (!status.ok()) {
|
|
|
|
Error(log_, "Error opening random access file %s. %s", Path().c_str(),
|
|
|
|
status.ToString().c_str());
|
|
|
|
return false;
|
|
|
|
}
|
Introduce a new storage specific Env API (#5761)
Summary:
The current Env API encompasses both storage/file operations, as well as OS related operations. Most of the APIs return a Status, which does not have enough metadata about an error, such as whether its retry-able or not, scope (i.e fault domain) of the error etc., that may be required in order to properly handle a storage error. The file APIs also do not provide enough control over the IO SLA, such as timeout, prioritization, hinting about placement and redundancy etc.
This PR separates out the file/storage APIs from Env into a new FileSystem class. The APIs are updated to return an IOStatus with metadata about the error, as well as to take an IOOptions structure as input in order to allow more control over the IO.
The user can set both ```options.env``` and ```options.file_system``` to specify that RocksDB should use the former for OS related operations and the latter for storage operations. Internally, a ```CompositeEnvWrapper``` has been introduced that inherits from ```Env``` and redirects individual methods to either an ```Env``` implementation or the ```FileSystem``` as appropriate. When options are sanitized during ```DB::Open```, ```options.env``` is replaced with a newly allocated ```CompositeEnvWrapper``` instance if both env and file_system have been specified. This way, the rest of the RocksDB code can continue to function as before.
This PR also ports PosixEnv to the new API by splitting it into two - PosixEnv and PosixFileSystem. PosixEnv is defined as a sub-class of CompositeEnvWrapper, and threading/time functions are overridden with Posix specific implementations in order to avoid an extra level of indirection.
The ```CompositeEnvWrapper``` translates ```IOStatus``` return code to ```Status```, and sets the severity to ```kSoftError``` if the io_status is retryable. The error handling code in RocksDB can then recover the DB automatically.
Pull Request resolved: https://github.com/facebook/rocksdb/pull/5761
Differential Revision: D18868376
Pulled By: anand1976
fbshipit-source-id: 39efe18a162ea746fabac6360ff529baba48486f
2019-12-13 22:47:08 +00:00
|
|
|
freader_.reset(new RandomAccessFileReader(
|
|
|
|
NewLegacyRandomAccessFileWrapper(file), Path(), env_));
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool RandomAccessCacheFile::Read(const LBA& lba, Slice* key, Slice* val,
|
|
|
|
char* scratch) {
|
|
|
|
ReadLock _(&rwlock_);
|
|
|
|
|
|
|
|
assert(lba.cache_id_ == cache_id_);
|
2016-11-22 18:26:08 +00:00
|
|
|
|
2017-02-27 20:59:34 +00:00
|
|
|
if (!freader_) {
|
2016-11-22 18:26:08 +00:00
|
|
|
return false;
|
|
|
|
}
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
Slice result;
|
2020-04-30 21:48:51 +00:00
|
|
|
Status s = freader_->Read(IOOptions(), lba.off_, lba.size_, &result, scratch,
|
|
|
|
nullptr);
|
2016-04-22 23:18:33 +00:00
|
|
|
if (!s.ok()) {
|
|
|
|
Error(log_, "Error reading from file %s. %s", Path().c_str(),
|
|
|
|
s.ToString().c_str());
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(result.data() == scratch);
|
|
|
|
|
|
|
|
return ParseRec(lba, key, val, scratch);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool RandomAccessCacheFile::ParseRec(const LBA& lba, Slice* key, Slice* val,
|
|
|
|
char* scratch) {
|
|
|
|
Slice data(scratch, lba.size_);
|
|
|
|
|
|
|
|
CacheRecord rec;
|
|
|
|
if (!rec.Deserialize(data)) {
|
|
|
|
assert(!"Error deserializing data");
|
|
|
|
Error(log_, "Error de-serializing record from file %s off %d",
|
|
|
|
Path().c_str(), lba.off_);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
*key = Slice(rec.key_);
|
|
|
|
*val = Slice(rec.val_);
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// WriteableCacheFile
|
|
|
|
//
|
|
|
|
|
|
|
|
WriteableCacheFile::~WriteableCacheFile() {
|
|
|
|
WriteLock _(&rwlock_);
|
|
|
|
if (!eof_) {
|
|
|
|
// This file never flushed. We give priority to shutdown since this is a
|
|
|
|
// cache
|
|
|
|
// TODO(krad): Figure a way to flush the pending data
|
2016-11-22 18:26:08 +00:00
|
|
|
if (file_) {
|
|
|
|
assert(refs_ == 1);
|
|
|
|
--refs_;
|
|
|
|
}
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
2016-11-22 18:26:08 +00:00
|
|
|
assert(!refs_);
|
2016-04-22 23:18:33 +00:00
|
|
|
ClearBuffers();
|
|
|
|
}
|
|
|
|
|
2018-03-05 21:08:17 +00:00
|
|
|
bool WriteableCacheFile::Create(const bool /*enable_direct_writes*/,
|
2016-05-03 18:05:42 +00:00
|
|
|
const bool enable_direct_reads) {
|
2016-04-22 23:18:33 +00:00
|
|
|
WriteLock _(&rwlock_);
|
|
|
|
|
2016-05-03 18:05:42 +00:00
|
|
|
enable_direct_reads_ = enable_direct_reads;
|
|
|
|
|
2017-03-16 02:22:52 +00:00
|
|
|
ROCKS_LOG_DEBUG(log_, "Creating new cache %s (max size is %d B)",
|
|
|
|
Path().c_str(), max_size_);
|
2016-04-22 23:18:33 +00:00
|
|
|
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(env_);
|
|
|
|
|
2016-04-22 23:18:33 +00:00
|
|
|
Status s = env_->FileExists(Path());
|
|
|
|
if (s.ok()) {
|
2017-03-16 02:22:52 +00:00
|
|
|
ROCKS_LOG_WARN(log_, "File %s already exists. %s", Path().c_str(),
|
|
|
|
s.ToString().c_str());
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
s = NewWritableCacheFile(env_, Path(), &file_);
|
|
|
|
if (!s.ok()) {
|
2017-03-16 02:22:52 +00:00
|
|
|
ROCKS_LOG_WARN(log_, "Unable to create file %s. %s", Path().c_str(),
|
|
|
|
s.ToString().c_str());
|
2016-04-22 23:18:33 +00:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(!refs_);
|
|
|
|
++refs_;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool WriteableCacheFile::Append(const Slice& key, const Slice& val, LBA* lba) {
|
|
|
|
WriteLock _(&rwlock_);
|
|
|
|
|
|
|
|
if (eof_) {
|
|
|
|
// We can't append since the file is full
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// estimate the space required to store the (key, val)
|
|
|
|
uint32_t rec_size = CacheRecord::CalcSize(key, val);
|
|
|
|
|
|
|
|
if (!ExpandBuffer(rec_size)) {
|
|
|
|
// unable to expand the buffer
|
2017-03-16 02:22:52 +00:00
|
|
|
ROCKS_LOG_DEBUG(log_, "Error expanding buffers. size=%d", rec_size);
|
2016-04-22 23:18:33 +00:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
lba->cache_id_ = cache_id_;
|
|
|
|
lba->off_ = disk_woff_;
|
|
|
|
lba->size_ = rec_size;
|
|
|
|
|
|
|
|
CacheRecord rec(key, val);
|
|
|
|
if (!rec.Serialize(&bufs_, &buf_woff_)) {
|
|
|
|
// unexpected error: unable to serialize the data
|
|
|
|
assert(!"Error serializing record");
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
disk_woff_ += rec_size;
|
|
|
|
eof_ = disk_woff_ >= max_size_;
|
|
|
|
|
|
|
|
// dispatch buffer for flush
|
|
|
|
DispatchBuffer();
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool WriteableCacheFile::ExpandBuffer(const size_t size) {
|
|
|
|
rwlock_.AssertHeld();
|
|
|
|
assert(!eof_);
|
|
|
|
|
|
|
|
// determine if there is enough space
|
|
|
|
size_t free = 0; // compute the free space left in buffer
|
|
|
|
for (size_t i = buf_woff_; i < bufs_.size(); ++i) {
|
|
|
|
free += bufs_[i]->Free();
|
|
|
|
if (size <= free) {
|
|
|
|
// we have enough space in the buffer
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// expand the buffer until there is enough space to write `size` bytes
|
|
|
|
assert(free < size);
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(alloc_);
|
|
|
|
|
2016-04-22 23:18:33 +00:00
|
|
|
while (free < size) {
|
|
|
|
CacheWriteBuffer* const buf = alloc_->Allocate();
|
|
|
|
if (!buf) {
|
2017-03-16 02:22:52 +00:00
|
|
|
ROCKS_LOG_DEBUG(log_, "Unable to allocate buffers");
|
2016-04-22 23:18:33 +00:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2016-08-03 00:15:18 +00:00
|
|
|
size_ += static_cast<uint32_t>(buf->Free());
|
2016-04-22 23:18:33 +00:00
|
|
|
free += buf->Free();
|
|
|
|
bufs_.push_back(buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(free >= size);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
void WriteableCacheFile::DispatchBuffer() {
|
|
|
|
rwlock_.AssertHeld();
|
|
|
|
|
|
|
|
assert(bufs_.size());
|
|
|
|
assert(buf_doff_ <= buf_woff_);
|
|
|
|
assert(buf_woff_ <= bufs_.size());
|
|
|
|
|
|
|
|
if (pending_ios_) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!eof_ && buf_doff_ == buf_woff_) {
|
|
|
|
// dispatch buffer is pointing to write buffer and we haven't hit eof
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(eof_ || buf_doff_ < buf_woff_);
|
|
|
|
assert(buf_doff_ < bufs_.size());
|
|
|
|
assert(file_);
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(alloc_);
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
auto* buf = bufs_[buf_doff_];
|
|
|
|
const uint64_t file_off = buf_doff_ * alloc_->BufferSize();
|
|
|
|
|
|
|
|
assert(!buf->Free() ||
|
|
|
|
(eof_ && buf_doff_ == buf_woff_ && buf_woff_ < bufs_.size()));
|
|
|
|
// we have reached end of file, and there is space in the last buffer
|
|
|
|
// pad it with zero for direct IO
|
|
|
|
buf->FillTrailingZeros();
|
|
|
|
|
2016-05-03 18:05:42 +00:00
|
|
|
assert(buf->Used() % kFileAlignmentSize == 0);
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
writer_->Write(file_.get(), buf, file_off,
|
|
|
|
std::bind(&WriteableCacheFile::BufferWriteDone, this));
|
|
|
|
pending_ios_++;
|
|
|
|
buf_doff_++;
|
|
|
|
}
|
|
|
|
|
|
|
|
void WriteableCacheFile::BufferWriteDone() {
|
|
|
|
WriteLock _(&rwlock_);
|
|
|
|
|
|
|
|
assert(bufs_.size());
|
|
|
|
|
|
|
|
pending_ios_--;
|
|
|
|
|
|
|
|
if (buf_doff_ < bufs_.size()) {
|
|
|
|
DispatchBuffer();
|
|
|
|
}
|
|
|
|
|
|
|
|
if (eof_ && buf_doff_ >= bufs_.size() && !pending_ios_) {
|
|
|
|
// end-of-file reached, move to read mode
|
|
|
|
CloseAndOpenForReading();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void WriteableCacheFile::CloseAndOpenForReading() {
|
|
|
|
// Our env abstraction do not allow reading from a file opened for appending
|
|
|
|
// We need close the file and re-open it for reading
|
|
|
|
Close();
|
2016-05-03 18:05:42 +00:00
|
|
|
RandomAccessCacheFile::OpenImpl(enable_direct_reads_);
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
bool WriteableCacheFile::ReadBuffer(const LBA& lba, Slice* key, Slice* block,
|
|
|
|
char* scratch) {
|
|
|
|
rwlock_.AssertHeld();
|
|
|
|
|
|
|
|
if (!ReadBuffer(lba, scratch)) {
|
|
|
|
Error(log_, "Error reading from buffer. cache=%d off=%d", cache_id_,
|
|
|
|
lba.off_);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ParseRec(lba, key, block, scratch);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool WriteableCacheFile::ReadBuffer(const LBA& lba, char* data) {
|
|
|
|
rwlock_.AssertHeld();
|
|
|
|
|
|
|
|
assert(lba.off_ < disk_woff_);
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(alloc_);
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
// we read from the buffers like reading from a flat file. The list of buffers
|
|
|
|
// are treated as contiguous stream of data
|
|
|
|
|
|
|
|
char* tmp = data;
|
|
|
|
size_t pending_nbytes = lba.size_;
|
|
|
|
// start buffer
|
|
|
|
size_t start_idx = lba.off_ / alloc_->BufferSize();
|
|
|
|
// offset into the start buffer
|
|
|
|
size_t start_off = lba.off_ % alloc_->BufferSize();
|
|
|
|
|
|
|
|
assert(start_idx <= buf_woff_);
|
|
|
|
|
|
|
|
for (size_t i = start_idx; pending_nbytes && i < bufs_.size(); ++i) {
|
|
|
|
assert(i <= buf_woff_);
|
|
|
|
auto* buf = bufs_[i];
|
|
|
|
assert(i == buf_woff_ || !buf->Free());
|
|
|
|
// bytes to write to the buffer
|
|
|
|
size_t nbytes = pending_nbytes > (buf->Used() - start_off)
|
|
|
|
? (buf->Used() - start_off)
|
|
|
|
: pending_nbytes;
|
|
|
|
memcpy(tmp, buf->Data() + start_off, nbytes);
|
|
|
|
|
|
|
|
// left over to be written
|
|
|
|
pending_nbytes -= nbytes;
|
|
|
|
start_off = 0;
|
|
|
|
tmp += nbytes;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(!pending_nbytes);
|
|
|
|
if (pending_nbytes) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(tmp == data + lba.size_);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
void WriteableCacheFile::Close() {
|
|
|
|
rwlock_.AssertHeld();
|
|
|
|
|
|
|
|
assert(size_ >= max_size_);
|
|
|
|
assert(disk_woff_ >= max_size_);
|
|
|
|
assert(buf_doff_ == bufs_.size());
|
|
|
|
assert(bufs_.size() - buf_woff_ <= 1);
|
|
|
|
assert(!pending_ios_);
|
|
|
|
|
|
|
|
Info(log_, "Closing file %s. size=%d written=%d", Path().c_str(), size_,
|
|
|
|
disk_woff_);
|
|
|
|
|
|
|
|
ClearBuffers();
|
|
|
|
file_.reset();
|
|
|
|
|
|
|
|
assert(refs_);
|
|
|
|
--refs_;
|
|
|
|
}
|
|
|
|
|
|
|
|
void WriteableCacheFile::ClearBuffers() {
|
2019-09-18 22:22:46 +00:00
|
|
|
assert(alloc_);
|
|
|
|
|
2016-04-22 23:18:33 +00:00
|
|
|
for (size_t i = 0; i < bufs_.size(); ++i) {
|
|
|
|
alloc_->Deallocate(bufs_[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
bufs_.clear();
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// ThreadedFileWriter implementation
|
|
|
|
//
|
|
|
|
ThreadedWriter::ThreadedWriter(PersistentCacheTier* const cache,
|
|
|
|
const size_t qdepth, const size_t io_size)
|
|
|
|
: Writer(cache), io_size_(io_size) {
|
|
|
|
for (size_t i = 0; i < qdepth; ++i) {
|
2017-02-06 22:43:55 +00:00
|
|
|
port::Thread th(&ThreadedWriter::ThreadMain, this);
|
2016-04-22 23:18:33 +00:00
|
|
|
threads_.push_back(std::move(th));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void ThreadedWriter::Stop() {
|
|
|
|
// notify all threads to exit
|
|
|
|
for (size_t i = 0; i < threads_.size(); ++i) {
|
|
|
|
q_.Push(IO(/*signal=*/true));
|
|
|
|
}
|
|
|
|
|
|
|
|
// wait for all threads to exit
|
|
|
|
for (auto& th : threads_) {
|
|
|
|
th.join();
|
2016-05-03 18:05:42 +00:00
|
|
|
assert(!th.joinable());
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
2016-05-03 18:05:42 +00:00
|
|
|
threads_.clear();
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void ThreadedWriter::Write(WritableFile* const file, CacheWriteBuffer* buf,
|
|
|
|
const uint64_t file_off,
|
|
|
|
const std::function<void()> callback) {
|
|
|
|
q_.Push(IO(file, buf, file_off, callback));
|
|
|
|
}
|
|
|
|
|
|
|
|
void ThreadedWriter::ThreadMain() {
|
|
|
|
while (true) {
|
|
|
|
// Fetch the IO to process
|
|
|
|
IO io(q_.Pop());
|
|
|
|
if (io.signal_) {
|
|
|
|
// that's secret signal to exit
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Reserve space for writing the buffer
|
|
|
|
while (!cache_->Reserve(io.buf_->Used())) {
|
|
|
|
// We can fail to reserve space if every file in the system
|
|
|
|
// is being currently accessed
|
2016-08-03 00:15:18 +00:00
|
|
|
/* sleep override */
|
|
|
|
Env::Default()->SleepForMicroseconds(1000000);
|
2016-04-22 23:18:33 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
DispatchIO(io);
|
|
|
|
|
|
|
|
io.callback_();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void ThreadedWriter::DispatchIO(const IO& io) {
|
|
|
|
size_t written = 0;
|
|
|
|
while (written < io.buf_->Used()) {
|
|
|
|
Slice data(io.buf_->Data() + written, io_size_);
|
|
|
|
Status s = io.file_->Append(data);
|
|
|
|
assert(s.ok());
|
|
|
|
if (!s.ok()) {
|
|
|
|
// That is definite IO error to device. There is not much we can
|
|
|
|
// do but ignore the failure. This can lead to corruption of data on
|
|
|
|
// disk, but the cache will skip while reading
|
|
|
|
fprintf(stderr, "Error writing data to file. %s\n", s.ToString().c_str());
|
|
|
|
}
|
|
|
|
written += io_size_;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-20 20:07:53 +00:00
|
|
|
} // namespace ROCKSDB_NAMESPACE
|
2016-04-22 23:18:33 +00:00
|
|
|
|
|
|
|
#endif
|