mirror of
https://github.com/Xahau/xahaud.git
synced 2025-12-06 17:27:52 +00:00
1fdd726 Hotfix RocksDB 3.5 d67500a Add `make install` to Makefile in 3.5.fb. 4cb631a update HISTORY.md cfd0946 comments about the BlockBasedTableOptions migration in Options REVERT: 25888ae Merge pull request #329 from fyrz/master REVERT: 89833e5 Fixed signed-unsigned comparison warning in db_test.cc REVERT: fcac705 Fixed compile warning on Mac caused by unused variables. REVERT: b3343fd resolution for java build problem introduced by 5ec53f3edf62bec1b690ce12fb21a6c52203f3c8 REVERT: 187b299 ForwardIterator: update prev_key_ only if prefix hasn't changed REVERT: 5ec53f3 make compaction related options changeable REVERT: d122e7b Update INSTALL.md REVERT: 986dad0 Merge pull request #324 from dalgaaf/wip-da-SCA-20140930 REVERT: 8ee75dc db/memtable.cc: remove unused variable merge_result REVERT: 0fd8bbc db/db_impl.cc: reduce scope of prefix_initialized REVERT: 676ff7b compaction_picker.cc: remove check for >=0 for unsigned REVERT: e55aea5 document_db.cc: fix assert REVERT: d517c83 in_table_factory.cc: use correct format specifier REVERT: b140375 ttl/ttl_test.cc: prefer prefix ++operator for non-primitive types REVERT: 43c789c spatialdb/spatial_db.cc: use !empty() instead of 'size() > 0' REVERT: 0de452e document_db.cc: pass const parameter by reference REVERT: 4cc8643 util/ldb_cmd.cc: prefer prefix ++operator for non-primitive types REVERT: af8c2b2 util/signal_test.cc: suppress intentional null pointer deref REVERT: 33580fa db/db_impl.cc: fix object handling, remove double lines REVERT: 873f135 db_ttl_impl.h: pass func parameter by reference REVERT: 8558457 ldb_cmd_execute_result.h: perform init in initialization list REVERT: 063471b table/table_test.cc: pass func parameter by reference REVERT: 93548ce table/cuckoo_table_reader.cc: pass func parameter by ref REVERT: b8b7117 db/version_set.cc: use !empty() instead of 'size() > 0' REVERT: 8ce050b table/bloom_block.*: pass func parameter by reference REVERT: 53910dd db_test.cc: pass parameter by reference REVERT: 68ca534 corruption_test.cc: pass parameter by reference REVERT: 7506198 cuckoo_table_db_test.cc: add flush after delete REVERT: 1f96330 Print MB per second compaction throughput separately for reads and writes REVERT: ffe3d49 Add an instruction about SSE in INSTALL.md REVERT: ee1f3cc Package generation for Ubuntu and CentOS REVERT: f0f7955 Fixing comile errors on OS X REVERT: 99fb613 remove 2 space linter REVERT: b2d64a4 Fix linters, second try REVERT: 747523d Print per column family metrics in db_bench REVERT: 56ebd40 Fix arc lint (should fix #238) REVERT: 637f891 Merge pull request #321 from eonnen/master REVERT: 827e31c Make test use a compatible type in the size checks. REVERT: fd5d80d CompactedDB: log using the correct info_log REVERT: 2faf49d use GetContext to replace callback function pointer REVERT: 983d2de Add AUTHORS file. Fix #203 REVERT: abd70c5 Merge pull request #316 from fyrz/ReverseBytewiseComparator REVERT: 2dc6f62 handle kDelete type in cuckoo builder REVERT: 8b8011a Changed name of ReverseBytewiseComparator based on review comment REVERT: 389edb6 universal compaction picker: use double for potential overflow REVERT: 5340484 Built-in comparator(s) in RocksJava REVERT: d439451 delay initialization of cuckoo table iterator REVERT: 94997ea reduce memory usage of cuckoo table builder REVERT: c627595 improve memory efficiency of cuckoo reader REVERT: 581442d option to choose module when calculating CuckooTable hash REVERT: fbd2daf CompactedDBImpl::MultiGet() for better CuckooTable performance REVERT: 3c68006 CompactedDBImpl REVERT: f7375f3 Fix double deletes REVERT: 21ddcf6 Remove allow_thread_local REVERT: fb4a492 Merge pull request #311 from ankgup87/master REVERT: 611e286 Merge branch 'master' of https://github.com/facebook/rocksdb REVERT: 0103b44 Merge branch 'master' of ssh://github.com/ankgup87/rocksdb REVERT: 1dfb7bb Add block based table config options REVERT: cdaf44f Enlarge log size cap when printing file summary REVERT: 7cc1ed7 Merge pull request #309 from naveenatceg/staticbuild REVERT: ba6d660 Resolving merge conflict REVERT: 51eeaf6 Addressing review comments REVERT: fd7d3fe Addressing review comments (adding a env variable to override temp directory) REVERT: cf7ace8 Addressing review comments REVERT: 0a29ce5 re-enable BlockBasedTable::SetupForCompaction() REVERT: 55af370 Remove TODO for checking index checksums REVERT: 3d74f09 Fix compile REVERT: 53b0039 Fix release compile REVERT: d0de413 WriteBatchWithIndex to allow different Comparators for different column families REVERT: 57a32f1 change target_file_size_base to uint64_t REVERT: 5e6aee4 dont create backup_input if compaction filter v2 is not used REVERT: 49b5f94 Merge pull request #306 from Liuchang0812/fix_cast REVERT: 787cb4d remove cast, replace %llu with % PRIu64 REVERT: a7574d4 Update logging.cc REVERT: 7e0dcb9 Update logging.cc REVERT: 57fa3cc Merge pull request #304 from Liuchang0812/fix-check REVERT: cd44522 Merge pull request #305 from Liuchang0812/fix-logging REVERT: 6a031b6 remove unused variable REVERT: 4436f17 fixed #303: replace %ld with % PRId64 REVERT: 7a1bd05 Merge pull request #302 from ankgup87/master REVERT: 423e52c Merge branch 'master' of https://github.com/facebook/rocksdb REVERT: bfeef94 Add rate limiter REVERT: 32f2532 Print compression_size_percent as a signed int REVERT: 976caca Skip AllocateTest if fallocate() is not supported in the file system REVERT: 3b897cd Enable no-fbcode RocksDB build REVERT: f445947 RocksDB: Format uint64 using PRIu64 in db_impl.cc REVERT: e17bc65 Merge pull request #299 from ankgup87/master REVERT: b93797a Fix build REVERT: adae3ca [Java] Fix JNI link error caused by the removal of options.db_stats_log_interval REVERT: 90b8c07 Fix unit tests errors REVERT: 51af7c3 CuckooTable: add one option to allow identity function for the first hash function REVERT: 0350435 Fixed a signed-unsigned comparison in spatial_db.cc -- issue #293 REVERT: 2fb1fea Fix syncronization issues REVERT: ff76895 Remove some unnecessary constructors REVERT: feadb9d fix cuckoo table builder test REVERT: 3c232e1 Fix mac compile REVERT: 54cada9 Run make format on PR #249 REVERT: 27b22f1 Merge pull request #249 from tdfischer/decompression-refactoring REVERT: fb6456b Replace naked calls to operator new and delete (Fixes #222) REVERT: 5600c8f cuckoo table: return estimated size - 1 REVERT: a062e1f SetOptions() for memtable related options REVERT: e4eca6a Options conversion function for convenience REVERT: a7c2094 Merge pull request #292 from saghmrossi/master REVERT: 4d05234 Merge branch 'master' of github.com:saghmrossi/rocksdb REVERT: 60a4aa1 Test use_mmap_reads REVERT: 94e43a1 [Java] Fixed 32-bit overflowing issue when converting jlong to size_t REVERT: f9eaaa6 added include for inttypes.h to fix nonworking printf statements REVERT: f090575 Replaced "built on on earlier work" by "built on earlier work" in README.md REVERT: faad439 Fix #284 REVERT: 49aacd8 Fix make install REVERT: acb9348 [Java] Include WriteBatch into RocksDBSample.java, fix how DbBenchmark.java handles WriteBatch. REVERT: 4a27a2f Don't sync manifest when disableDataSync = true REVERT: 9b8480d Merge pull request #287 from yinqiwen/rate-limiter-crash-fix REVERT: 28be16b fix rate limiter crash #286 REVERT: 04ce1b2 Fix #284 REVERT: add22e3 standardize scripts to run RocksDB benchmarks REVERT: dee91c2 WriteThread REVERT: 540a257 Fix WAL synced REVERT: 24f034b Merge pull request #282 from Chilledheart/develop REVERT: 49fe329 Fix build issue under macosx REVERT: ebb5c65 Add make install REVERT: 0352a9f add_wrapped_bloom_test REVERT: 9c0e66c Don't run background jobs (flush, compactions) when bg_error_ is set REVERT: a9639bd Fix valgrind test REVERT: d1f24dc Relax FlushSchedule test REVERT: 3d9e6f7 Push model for flushing memtables REVERT: 059e584 [unit test] CompactRange should fail if we don't have space REVERT: dd641b2 fix RocksDB java build REVERT: 53404d9 add_qps_info_in cache bench REVERT: a52cecb Fix Mac compile REVERT: 092f97e Fix comments and typos REVERT: 6cc1286 Added a few statistics for BackupableDB REVERT: 0a42295 Fix SimpleWriteTimeoutTest REVERT: 06d9862 Always pass MergeContext as pointer, not reference REVERT: d343c3f Improve db recovery REVERT: 6bb7e3e Merger test REVERT: 88841bd Explicitly cast char to signed char in Hash() REVERT: 5231146 MemTableOptions REVERT: 1d284db Addressing review comments REVERT: 55114e7 Some updates for SpatialDB REVERT: 171d4ff remove TailingIterator reference in db_impl.h REVERT: 9b0f7ff rename version_set options_ to db_options_ to avoid confusion REVERT: 2d57828 Check stop level trigger-0 before slowdown level-0 trigger REVERT: 659d2d5 move compaction_filter to immutable_options REVERT: 048560a reduce references to cfd->options() in DBImpl REVERT: 011241b DB::Flush() Do not wait for background threads when there is nothing in mem table REVERT: a2bb7c3 Push- instead of pull-model for managing Write stalls REVERT: 0af157f Implement full filter for block based table. REVERT: 9360cc6 Fix valgrind issue REVERT: 02d5bff Merge pull request #277 from wankai/master REVERT: 88a2f44 fix comments REVERT: 7c16e39 Merge pull request #276 from wankai/master REVERT: 8237738 replace hard-coded number with named variable REVERT: db8ca52 Merge pull request #273 from nbougalis/static-analysis REVERT: b7b031f Merge pull request #274 from wankai/master REVERT: 4c2b1f0 Merge remote-tracking branch 'upstream/master' REVERT: a5d2863 typo improvement REVERT: 9f8aa09 Don't leak data returned by opendir REVERT: d1cfb71 Remove unused member(s) REVERT: bfee319 sizeof(int*) where sizeof(int) was intended REVERT: d40c1f7 Add missing break statement REVERT: 2e97c38 Avoid off-by-one error when using readlink REVERT: 40ddc3d add cache bench REVERT: 9f1c80b Drop column family from write thread REVERT: 8de151b Add db_bench with lots of column families to regression tests REVERT: c9e419c rename options_ to db_options_ in DBImpl to avoid confusion REVERT: 5cd0576 Fix compaction bug in Cuckoo Table Builder. Use kvs_.size() instead of num_entries in FileSize() method. REVERT: 0fbb3fa fixed memory leak in unit test DBIteratorBoundTest REVERT: adcd253 fix asan check REVERT: 4092b7a Merge pull request #272 from project-zerus/patch-1 REVERT: bb6ae0f fix more compile warnings REVERT: 6d31441 Merge pull request #271 from nbougalis/cleanups REVERT: 0cd0ec4 Plug memory leak during index creation REVERT: 4329d74 Fix swapped variable names to accurately reflect usage REVERT: 45a5e3e Remove path with arena==nullptr from NewInternalIterator REVERT: 5665e5e introduce ImmutableOptions REVERT: e0b99d4 created a new ReadOptions parameter 'iterate_upper_bound' REVERT: 51ea889 Fix travis builds REVERT: a481626 Relax backupable rate limiting test REVERT: f7f973d Merge pull request #269 from huahang/patch-2 REVERT: ef5b384 fix a few compile warnings REVERT: 2fd3806 Merge pull request #263 from wankai/master REVERT: 1785114 delete unused Comparator REVERT: 1b1d961 update HISTORY.md REVERT: 703c3ea comments about the BlockBasedTableOptions migration in Options REVERT: 4b5ad88 Merge pull request #260 from wankai/master REVERT: 19cc588 change to filter_block std::unique_ptr support RAII REVERT: 9b976e3 Merge pull request #259 from wankai/master REVERT: 5d25a46 Merge remote-tracking branch 'upstream/master' REVERT: dff2b1a typo improvement REVERT: 343e98a Reverting import change REVERT: ddb8039 RocksDB static build Make file changes to download and build the dependencies .Load the shared library when RocksDB is initialized git-subtree-dir: src/rocksdb2 git-subtree-split: 1fdd726a8254c13d0c66d8db8130ad17c13d7bcc
1118 lines
37 KiB
C++
1118 lines
37 KiB
C++
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
|
// This source code is licensed under the BSD-style license found in the
|
|
// LICENSE file in the root directory of this source tree. An additional grant
|
|
// of patent rights can be found in the PATENTS file in the same directory.
|
|
//
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
#include <algorithm>
|
|
#include <set>
|
|
|
|
#include "db/db_impl.h"
|
|
#include "db/filename.h"
|
|
#include "db/version_set.h"
|
|
#include "db/write_batch_internal.h"
|
|
#include "rocksdb/cache.h"
|
|
#include "rocksdb/compaction_filter.h"
|
|
#include "rocksdb/db.h"
|
|
#include "rocksdb/env.h"
|
|
#include "rocksdb/filter_policy.h"
|
|
#include "rocksdb/slice_transform.h"
|
|
#include "rocksdb/table.h"
|
|
#include "table/meta_blocks.h"
|
|
#include "table/bloom_block.h"
|
|
#include "table/plain_table_factory.h"
|
|
#include "table/plain_table_reader.h"
|
|
#include "util/hash.h"
|
|
#include "util/logging.h"
|
|
#include "util/mutexlock.h"
|
|
#include "util/testharness.h"
|
|
#include "util/testutil.h"
|
|
#include "utilities/merge_operators.h"
|
|
|
|
using std::unique_ptr;
|
|
|
|
namespace rocksdb {
|
|
|
|
class PlainTableDBTest {
|
|
protected:
|
|
private:
|
|
std::string dbname_;
|
|
Env* env_;
|
|
DB* db_;
|
|
|
|
Options last_options_;
|
|
|
|
public:
|
|
PlainTableDBTest() : env_(Env::Default()) {
|
|
dbname_ = test::TmpDir() + "/plain_table_db_test";
|
|
ASSERT_OK(DestroyDB(dbname_, Options()));
|
|
db_ = nullptr;
|
|
Reopen();
|
|
}
|
|
|
|
~PlainTableDBTest() {
|
|
delete db_;
|
|
ASSERT_OK(DestroyDB(dbname_, Options()));
|
|
}
|
|
|
|
// Return the current option configuration.
|
|
Options CurrentOptions() {
|
|
Options options;
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 0;
|
|
plain_table_options.bloom_bits_per_key = 2;
|
|
plain_table_options.hash_table_ratio = 0.8;
|
|
plain_table_options.index_sparseness = 3;
|
|
plain_table_options.huge_page_tlb_size = 0;
|
|
plain_table_options.encoding_type = kPrefix;
|
|
plain_table_options.full_scan_mode = false;
|
|
plain_table_options.store_index_in_file = false;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory(plain_table_options));
|
|
options.memtable_factory.reset(NewHashLinkListRepFactory(4, 0, 3, true));
|
|
|
|
options.prefix_extractor.reset(NewFixedPrefixTransform(8));
|
|
options.allow_mmap_reads = true;
|
|
return options;
|
|
}
|
|
|
|
DBImpl* dbfull() {
|
|
return reinterpret_cast<DBImpl*>(db_);
|
|
}
|
|
|
|
void Reopen(Options* options = nullptr) {
|
|
ASSERT_OK(TryReopen(options));
|
|
}
|
|
|
|
void Close() {
|
|
delete db_;
|
|
db_ = nullptr;
|
|
}
|
|
|
|
void DestroyAndReopen(Options* options = nullptr) {
|
|
//Destroy using last options
|
|
Destroy(&last_options_);
|
|
ASSERT_OK(TryReopen(options));
|
|
}
|
|
|
|
void Destroy(Options* options) {
|
|
delete db_;
|
|
db_ = nullptr;
|
|
ASSERT_OK(DestroyDB(dbname_, *options));
|
|
}
|
|
|
|
Status PureReopen(Options* options, DB** db) {
|
|
return DB::Open(*options, dbname_, db);
|
|
}
|
|
|
|
Status TryReopen(Options* options = nullptr) {
|
|
delete db_;
|
|
db_ = nullptr;
|
|
Options opts;
|
|
if (options != nullptr) {
|
|
opts = *options;
|
|
} else {
|
|
opts = CurrentOptions();
|
|
opts.create_if_missing = true;
|
|
}
|
|
last_options_ = opts;
|
|
|
|
return DB::Open(opts, dbname_, &db_);
|
|
}
|
|
|
|
Status Put(const Slice& k, const Slice& v) {
|
|
return db_->Put(WriteOptions(), k, v);
|
|
}
|
|
|
|
Status Delete(const std::string& k) {
|
|
return db_->Delete(WriteOptions(), k);
|
|
}
|
|
|
|
std::string Get(const std::string& k, const Snapshot* snapshot = nullptr) {
|
|
ReadOptions options;
|
|
options.snapshot = snapshot;
|
|
std::string result;
|
|
Status s = db_->Get(options, k, &result);
|
|
if (s.IsNotFound()) {
|
|
result = "NOT_FOUND";
|
|
} else if (!s.ok()) {
|
|
result = s.ToString();
|
|
}
|
|
return result;
|
|
}
|
|
|
|
|
|
int NumTableFilesAtLevel(int level) {
|
|
std::string property;
|
|
ASSERT_TRUE(
|
|
db_->GetProperty("rocksdb.num-files-at-level" + NumberToString(level),
|
|
&property));
|
|
return atoi(property.c_str());
|
|
}
|
|
|
|
// Return spread of files per level
|
|
std::string FilesPerLevel() {
|
|
std::string result;
|
|
int last_non_zero_offset = 0;
|
|
for (int level = 0; level < db_->NumberLevels(); level++) {
|
|
int f = NumTableFilesAtLevel(level);
|
|
char buf[100];
|
|
snprintf(buf, sizeof(buf), "%s%d", (level ? "," : ""), f);
|
|
result += buf;
|
|
if (f > 0) {
|
|
last_non_zero_offset = result.size();
|
|
}
|
|
}
|
|
result.resize(last_non_zero_offset);
|
|
return result;
|
|
}
|
|
|
|
std::string IterStatus(Iterator* iter) {
|
|
std::string result;
|
|
if (iter->Valid()) {
|
|
result = iter->key().ToString() + "->" + iter->value().ToString();
|
|
} else {
|
|
result = "(invalid)";
|
|
}
|
|
return result;
|
|
}
|
|
};
|
|
|
|
TEST(PlainTableDBTest, Empty) {
|
|
ASSERT_TRUE(dbfull() != nullptr);
|
|
ASSERT_EQ("NOT_FOUND", Get("0000000000000foo"));
|
|
}
|
|
|
|
extern const uint64_t kPlainTableMagicNumber;
|
|
|
|
class TestPlainTableReader : public PlainTableReader {
|
|
public:
|
|
TestPlainTableReader(const EnvOptions& storage_options,
|
|
const InternalKeyComparator& icomparator,
|
|
EncodingType encoding_type, uint64_t file_size,
|
|
int bloom_bits_per_key, double hash_table_ratio,
|
|
size_t index_sparseness,
|
|
const TableProperties* table_properties,
|
|
unique_ptr<RandomAccessFile>&& file,
|
|
const Options& options, bool* expect_bloom_not_match,
|
|
bool store_index_in_file)
|
|
: PlainTableReader(options, std::move(file), storage_options, icomparator,
|
|
encoding_type, file_size, table_properties),
|
|
expect_bloom_not_match_(expect_bloom_not_match) {
|
|
Status s = MmapDataFile();
|
|
ASSERT_TRUE(s.ok());
|
|
|
|
s = PopulateIndex(const_cast<TableProperties*>(table_properties),
|
|
bloom_bits_per_key, hash_table_ratio, index_sparseness,
|
|
2 * 1024 * 1024);
|
|
ASSERT_TRUE(s.ok());
|
|
|
|
TableProperties* props = const_cast<TableProperties*>(table_properties);
|
|
if (store_index_in_file) {
|
|
auto bloom_version_ptr = props->user_collected_properties.find(
|
|
PlainTablePropertyNames::kBloomVersion);
|
|
ASSERT_TRUE(bloom_version_ptr != props->user_collected_properties.end());
|
|
ASSERT_EQ(bloom_version_ptr->second, std::string("1"));
|
|
if (options.bloom_locality > 0) {
|
|
auto num_blocks_ptr = props->user_collected_properties.find(
|
|
PlainTablePropertyNames::kNumBloomBlocks);
|
|
ASSERT_TRUE(num_blocks_ptr != props->user_collected_properties.end());
|
|
}
|
|
}
|
|
}
|
|
|
|
virtual ~TestPlainTableReader() {}
|
|
|
|
private:
|
|
virtual bool MatchBloom(uint32_t hash) const override {
|
|
bool ret = PlainTableReader::MatchBloom(hash);
|
|
if (*expect_bloom_not_match_) {
|
|
ASSERT_TRUE(!ret);
|
|
} else {
|
|
ASSERT_TRUE(ret);
|
|
}
|
|
return ret;
|
|
}
|
|
bool* expect_bloom_not_match_;
|
|
};
|
|
|
|
extern const uint64_t kPlainTableMagicNumber;
|
|
class TestPlainTableFactory : public PlainTableFactory {
|
|
public:
|
|
explicit TestPlainTableFactory(bool* expect_bloom_not_match,
|
|
const PlainTableOptions& options)
|
|
: PlainTableFactory(options),
|
|
bloom_bits_per_key_(options.bloom_bits_per_key),
|
|
hash_table_ratio_(options.hash_table_ratio),
|
|
index_sparseness_(options.index_sparseness),
|
|
store_index_in_file_(options.store_index_in_file),
|
|
expect_bloom_not_match_(expect_bloom_not_match) {}
|
|
|
|
Status NewTableReader(const Options& options, const EnvOptions& soptions,
|
|
const InternalKeyComparator& internal_comparator,
|
|
unique_ptr<RandomAccessFile>&& file, uint64_t file_size,
|
|
unique_ptr<TableReader>* table) const override {
|
|
TableProperties* props = nullptr;
|
|
auto s = ReadTableProperties(file.get(), file_size, kPlainTableMagicNumber,
|
|
options.env, options.info_log.get(), &props);
|
|
ASSERT_TRUE(s.ok());
|
|
|
|
if (store_index_in_file_) {
|
|
BlockHandle bloom_block_handle;
|
|
s = FindMetaBlock(file.get(), file_size, kPlainTableMagicNumber,
|
|
options.env, BloomBlockBuilder::kBloomBlock,
|
|
&bloom_block_handle);
|
|
ASSERT_TRUE(s.ok());
|
|
|
|
BlockHandle index_block_handle;
|
|
s = FindMetaBlock(
|
|
file.get(), file_size, kPlainTableMagicNumber, options.env,
|
|
PlainTableIndexBuilder::kPlainTableIndexBlock, &index_block_handle);
|
|
ASSERT_TRUE(s.ok());
|
|
}
|
|
|
|
auto& user_props = props->user_collected_properties;
|
|
auto encoding_type_prop =
|
|
user_props.find(PlainTablePropertyNames::kEncodingType);
|
|
assert(encoding_type_prop != user_props.end());
|
|
EncodingType encoding_type = static_cast<EncodingType>(
|
|
DecodeFixed32(encoding_type_prop->second.c_str()));
|
|
|
|
std::unique_ptr<PlainTableReader> new_reader(new TestPlainTableReader(
|
|
soptions, internal_comparator, encoding_type, file_size,
|
|
bloom_bits_per_key_, hash_table_ratio_, index_sparseness_, props,
|
|
std::move(file), options, expect_bloom_not_match_,
|
|
store_index_in_file_));
|
|
|
|
*table = std::move(new_reader);
|
|
return s;
|
|
}
|
|
|
|
private:
|
|
int bloom_bits_per_key_;
|
|
double hash_table_ratio_;
|
|
size_t index_sparseness_;
|
|
bool store_index_in_file_;
|
|
bool* expect_bloom_not_match_;
|
|
};
|
|
|
|
TEST(PlainTableDBTest, Flush) {
|
|
for (size_t huge_page_tlb_size = 0; huge_page_tlb_size <= 2 * 1024 * 1024;
|
|
huge_page_tlb_size += 2 * 1024 * 1024) {
|
|
for (EncodingType encoding_type : {kPlain, kPrefix}) {
|
|
for (int bloom_bits = 0; bloom_bits <= 117; bloom_bits += 117) {
|
|
for (int total_order = 0; total_order <= 1; total_order++) {
|
|
for (int store_index_in_file = 0; store_index_in_file <= 1;
|
|
++store_index_in_file) {
|
|
if (!bloom_bits && store_index_in_file) {
|
|
continue;
|
|
}
|
|
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
if (total_order) {
|
|
options.prefix_extractor.reset();
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 0;
|
|
plain_table_options.bloom_bits_per_key = bloom_bits;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
plain_table_options.index_sparseness = 2;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
plain_table_options.encoding_type = encoding_type;
|
|
plain_table_options.full_scan_mode = false;
|
|
plain_table_options.store_index_in_file = store_index_in_file;
|
|
|
|
options.table_factory.reset(
|
|
NewPlainTableFactory(plain_table_options));
|
|
} else {
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 0;
|
|
plain_table_options.bloom_bits_per_key = bloom_bits;
|
|
plain_table_options.hash_table_ratio = 0.75;
|
|
plain_table_options.index_sparseness = 16;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
plain_table_options.encoding_type = encoding_type;
|
|
plain_table_options.full_scan_mode = false;
|
|
plain_table_options.store_index_in_file = store_index_in_file;
|
|
|
|
options.table_factory.reset(
|
|
NewPlainTableFactory(plain_table_options));
|
|
}
|
|
DestroyAndReopen(&options);
|
|
uint64_t int_num;
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
"rocksdb.estimate-table-readers-mem", &int_num));
|
|
ASSERT_EQ(int_num, 0U);
|
|
|
|
ASSERT_OK(Put("1000000000000foo", "v1"));
|
|
ASSERT_OK(Put("0000000000000bar", "v2"));
|
|
ASSERT_OK(Put("1000000000000foo", "v3"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
ASSERT_TRUE(dbfull()->GetIntProperty(
|
|
"rocksdb.estimate-table-readers-mem", &int_num));
|
|
ASSERT_GT(int_num, 0U);
|
|
|
|
TablePropertiesCollection ptc;
|
|
reinterpret_cast<DB*>(dbfull())->GetPropertiesOfAllTables(&ptc);
|
|
ASSERT_EQ(1U, ptc.size());
|
|
auto row = ptc.begin();
|
|
auto tp = row->second;
|
|
|
|
if (!store_index_in_file) {
|
|
ASSERT_EQ(total_order ? "4" : "12",
|
|
(tp->user_collected_properties)
|
|
.at("plain_table_hash_table_size"));
|
|
ASSERT_EQ("0", (tp->user_collected_properties)
|
|
.at("plain_table_sub_index_size"));
|
|
} else {
|
|
ASSERT_EQ("0", (tp->user_collected_properties)
|
|
.at("plain_table_hash_table_size"));
|
|
ASSERT_EQ("0", (tp->user_collected_properties)
|
|
.at("plain_table_sub_index_size"));
|
|
}
|
|
ASSERT_EQ("v3", Get("1000000000000foo"));
|
|
ASSERT_EQ("v2", Get("0000000000000bar"));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(PlainTableDBTest, Flush2) {
|
|
for (size_t huge_page_tlb_size = 0; huge_page_tlb_size <= 2 * 1024 * 1024;
|
|
huge_page_tlb_size += 2 * 1024 * 1024) {
|
|
for (EncodingType encoding_type : {kPlain, kPrefix}) {
|
|
for (int bloom_bits = 0; bloom_bits <= 117; bloom_bits += 117) {
|
|
for (int total_order = 0; total_order <= 1; total_order++) {
|
|
for (int store_index_in_file = 0; store_index_in_file <= 1;
|
|
++store_index_in_file) {
|
|
if (encoding_type == kPrefix && total_order) {
|
|
continue;
|
|
}
|
|
if (!bloom_bits && store_index_in_file) {
|
|
continue;
|
|
}
|
|
if (total_order && store_index_in_file) {
|
|
continue;
|
|
}
|
|
bool expect_bloom_not_match = false;
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
PlainTableOptions plain_table_options;
|
|
if (total_order) {
|
|
options.prefix_extractor = nullptr;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
plain_table_options.index_sparseness = 2;
|
|
} else {
|
|
plain_table_options.hash_table_ratio = 0.75;
|
|
plain_table_options.index_sparseness = 16;
|
|
}
|
|
plain_table_options.user_key_len = kPlainTableVariableLength;
|
|
plain_table_options.bloom_bits_per_key = bloom_bits;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
plain_table_options.encoding_type = encoding_type;
|
|
plain_table_options.store_index_in_file = store_index_in_file;
|
|
options.table_factory.reset(new TestPlainTableFactory(
|
|
&expect_bloom_not_match, plain_table_options));
|
|
|
|
DestroyAndReopen(&options);
|
|
ASSERT_OK(Put("0000000000000bar", "b"));
|
|
ASSERT_OK(Put("1000000000000foo", "v1"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
ASSERT_OK(Put("1000000000000foo", "v2"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("v2", Get("1000000000000foo"));
|
|
|
|
ASSERT_OK(Put("0000000000000eee", "v3"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("v3", Get("0000000000000eee"));
|
|
|
|
ASSERT_OK(Delete("0000000000000bar"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("NOT_FOUND", Get("0000000000000bar"));
|
|
|
|
ASSERT_OK(Put("0000000000000eee", "v5"));
|
|
ASSERT_OK(Put("9000000000000eee", "v5"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("v5", Get("0000000000000eee"));
|
|
|
|
// Test Bloom Filter
|
|
if (bloom_bits > 0) {
|
|
// Neither key nor value should exist.
|
|
expect_bloom_not_match = true;
|
|
ASSERT_EQ("NOT_FOUND", Get("5_not00000000bar"));
|
|
// Key doesn't exist any more but prefix exists.
|
|
if (total_order) {
|
|
ASSERT_EQ("NOT_FOUND", Get("1000000000000not"));
|
|
ASSERT_EQ("NOT_FOUND", Get("0000000000000not"));
|
|
}
|
|
expect_bloom_not_match = false;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(PlainTableDBTest, Iterator) {
|
|
for (size_t huge_page_tlb_size = 0; huge_page_tlb_size <= 2 * 1024 * 1024;
|
|
huge_page_tlb_size += 2 * 1024 * 1024) {
|
|
for (EncodingType encoding_type : {kPlain, kPrefix}) {
|
|
for (int bloom_bits = 0; bloom_bits <= 117; bloom_bits += 117) {
|
|
for (int total_order = 0; total_order <= 1; total_order++) {
|
|
if (encoding_type == kPrefix && total_order == 1) {
|
|
continue;
|
|
}
|
|
bool expect_bloom_not_match = false;
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
if (total_order) {
|
|
options.prefix_extractor = nullptr;
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 16;
|
|
plain_table_options.bloom_bits_per_key = bloom_bits;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
plain_table_options.index_sparseness = 2;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
plain_table_options.encoding_type = encoding_type;
|
|
|
|
options.table_factory.reset(new TestPlainTableFactory(
|
|
&expect_bloom_not_match, plain_table_options));
|
|
} else {
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 16;
|
|
plain_table_options.bloom_bits_per_key = bloom_bits;
|
|
plain_table_options.hash_table_ratio = 0.75;
|
|
plain_table_options.index_sparseness = 16;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
plain_table_options.encoding_type = encoding_type;
|
|
|
|
options.table_factory.reset(new TestPlainTableFactory(
|
|
&expect_bloom_not_match, plain_table_options));
|
|
}
|
|
DestroyAndReopen(&options);
|
|
|
|
ASSERT_OK(Put("1000000000foo002", "v_2"));
|
|
ASSERT_OK(Put("0000000000000bar", "random"));
|
|
ASSERT_OK(Put("1000000000foo001", "v1"));
|
|
ASSERT_OK(Put("3000000000000bar", "bar_v"));
|
|
ASSERT_OK(Put("1000000000foo003", "v__3"));
|
|
ASSERT_OK(Put("1000000000foo004", "v__4"));
|
|
ASSERT_OK(Put("1000000000foo005", "v__5"));
|
|
ASSERT_OK(Put("1000000000foo007", "v__7"));
|
|
ASSERT_OK(Put("1000000000foo008", "v__8"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("v1", Get("1000000000foo001"));
|
|
ASSERT_EQ("v__3", Get("1000000000foo003"));
|
|
Iterator* iter = dbfull()->NewIterator(ReadOptions());
|
|
iter->Seek("1000000000foo000");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo001", iter->key().ToString());
|
|
ASSERT_EQ("v1", iter->value().ToString());
|
|
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo002", iter->key().ToString());
|
|
ASSERT_EQ("v_2", iter->value().ToString());
|
|
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo003", iter->key().ToString());
|
|
ASSERT_EQ("v__3", iter->value().ToString());
|
|
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo004", iter->key().ToString());
|
|
ASSERT_EQ("v__4", iter->value().ToString());
|
|
|
|
iter->Seek("3000000000000bar");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("3000000000000bar", iter->key().ToString());
|
|
ASSERT_EQ("bar_v", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo000");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo001", iter->key().ToString());
|
|
ASSERT_EQ("v1", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo005");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo005", iter->key().ToString());
|
|
ASSERT_EQ("v__5", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo006");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo007", iter->key().ToString());
|
|
ASSERT_EQ("v__7", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo008");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo008", iter->key().ToString());
|
|
ASSERT_EQ("v__8", iter->value().ToString());
|
|
|
|
if (total_order == 0) {
|
|
iter->Seek("1000000000foo009");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("3000000000000bar", iter->key().ToString());
|
|
}
|
|
|
|
// Test Bloom Filter
|
|
if (bloom_bits > 0) {
|
|
if (!total_order) {
|
|
// Neither key nor value should exist.
|
|
expect_bloom_not_match = true;
|
|
iter->Seek("2not000000000bar");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
ASSERT_EQ("NOT_FOUND", Get("2not000000000bar"));
|
|
expect_bloom_not_match = false;
|
|
} else {
|
|
expect_bloom_not_match = true;
|
|
ASSERT_EQ("NOT_FOUND", Get("2not000000000bar"));
|
|
expect_bloom_not_match = false;
|
|
}
|
|
}
|
|
|
|
delete iter;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
namespace {
|
|
std::string MakeLongKey(size_t length, char c) {
|
|
return std::string(length, c);
|
|
}
|
|
} // namespace
|
|
|
|
TEST(PlainTableDBTest, IteratorLargeKeys) {
|
|
Options options = CurrentOptions();
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 0;
|
|
plain_table_options.bloom_bits_per_key = 0;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory(plain_table_options));
|
|
options.create_if_missing = true;
|
|
options.prefix_extractor.reset();
|
|
DestroyAndReopen(&options);
|
|
|
|
std::string key_list[] = {
|
|
MakeLongKey(30, '0'),
|
|
MakeLongKey(16, '1'),
|
|
MakeLongKey(32, '2'),
|
|
MakeLongKey(60, '3'),
|
|
MakeLongKey(90, '4'),
|
|
MakeLongKey(50, '5'),
|
|
MakeLongKey(26, '6')
|
|
};
|
|
|
|
for (size_t i = 0; i < 7; i++) {
|
|
ASSERT_OK(Put(key_list[i], std::to_string(i)));
|
|
}
|
|
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
Iterator* iter = dbfull()->NewIterator(ReadOptions());
|
|
iter->Seek(key_list[0]);
|
|
|
|
for (size_t i = 0; i < 7; i++) {
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ(key_list[i], iter->key().ToString());
|
|
ASSERT_EQ(std::to_string(i), iter->value().ToString());
|
|
iter->Next();
|
|
}
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
delete iter;
|
|
}
|
|
|
|
namespace {
|
|
std::string MakeLongKeyWithPrefix(size_t length, char c) {
|
|
return "00000000" + std::string(length - 8, c);
|
|
}
|
|
} // namespace
|
|
|
|
TEST(PlainTableDBTest, IteratorLargeKeysWithPrefix) {
|
|
Options options = CurrentOptions();
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 16;
|
|
plain_table_options.bloom_bits_per_key = 0;
|
|
plain_table_options.hash_table_ratio = 0.8;
|
|
plain_table_options.index_sparseness = 3;
|
|
plain_table_options.huge_page_tlb_size = 0;
|
|
plain_table_options.encoding_type = kPrefix;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory(plain_table_options));
|
|
options.create_if_missing = true;
|
|
DestroyAndReopen(&options);
|
|
|
|
std::string key_list[] = {
|
|
MakeLongKeyWithPrefix(30, '0'), MakeLongKeyWithPrefix(16, '1'),
|
|
MakeLongKeyWithPrefix(32, '2'), MakeLongKeyWithPrefix(60, '3'),
|
|
MakeLongKeyWithPrefix(90, '4'), MakeLongKeyWithPrefix(50, '5'),
|
|
MakeLongKeyWithPrefix(26, '6')};
|
|
|
|
for (size_t i = 0; i < 7; i++) {
|
|
ASSERT_OK(Put(key_list[i], std::to_string(i)));
|
|
}
|
|
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
Iterator* iter = dbfull()->NewIterator(ReadOptions());
|
|
iter->Seek(key_list[0]);
|
|
|
|
for (size_t i = 0; i < 7; i++) {
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ(key_list[i], iter->key().ToString());
|
|
ASSERT_EQ(std::to_string(i), iter->value().ToString());
|
|
iter->Next();
|
|
}
|
|
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
delete iter;
|
|
}
|
|
|
|
// A test comparator which compare two strings in this way:
|
|
// (1) first compare prefix of 8 bytes in alphabet order,
|
|
// (2) if two strings share the same prefix, sort the other part of the string
|
|
// in the reverse alphabet order.
|
|
class SimpleSuffixReverseComparator : public Comparator {
|
|
public:
|
|
SimpleSuffixReverseComparator() {}
|
|
|
|
virtual const char* Name() const { return "SimpleSuffixReverseComparator"; }
|
|
|
|
virtual int Compare(const Slice& a, const Slice& b) const {
|
|
Slice prefix_a = Slice(a.data(), 8);
|
|
Slice prefix_b = Slice(b.data(), 8);
|
|
int prefix_comp = prefix_a.compare(prefix_b);
|
|
if (prefix_comp != 0) {
|
|
return prefix_comp;
|
|
} else {
|
|
Slice suffix_a = Slice(a.data() + 8, a.size() - 8);
|
|
Slice suffix_b = Slice(b.data() + 8, b.size() - 8);
|
|
return -(suffix_a.compare(suffix_b));
|
|
}
|
|
}
|
|
virtual void FindShortestSeparator(std::string* start,
|
|
const Slice& limit) const {}
|
|
|
|
virtual void FindShortSuccessor(std::string* key) const {}
|
|
};
|
|
|
|
TEST(PlainTableDBTest, IteratorReverseSuffixComparator) {
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
SimpleSuffixReverseComparator comp;
|
|
options.comparator = ∁
|
|
DestroyAndReopen(&options);
|
|
|
|
ASSERT_OK(Put("1000000000foo002", "v_2"));
|
|
ASSERT_OK(Put("0000000000000bar", "random"));
|
|
ASSERT_OK(Put("1000000000foo001", "v1"));
|
|
ASSERT_OK(Put("3000000000000bar", "bar_v"));
|
|
ASSERT_OK(Put("1000000000foo003", "v__3"));
|
|
ASSERT_OK(Put("1000000000foo004", "v__4"));
|
|
ASSERT_OK(Put("1000000000foo005", "v__5"));
|
|
ASSERT_OK(Put("1000000000foo007", "v__7"));
|
|
ASSERT_OK(Put("1000000000foo008", "v__8"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("v1", Get("1000000000foo001"));
|
|
ASSERT_EQ("v__3", Get("1000000000foo003"));
|
|
Iterator* iter = dbfull()->NewIterator(ReadOptions());
|
|
iter->Seek("1000000000foo009");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo008", iter->key().ToString());
|
|
ASSERT_EQ("v__8", iter->value().ToString());
|
|
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo007", iter->key().ToString());
|
|
ASSERT_EQ("v__7", iter->value().ToString());
|
|
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo005", iter->key().ToString());
|
|
ASSERT_EQ("v__5", iter->value().ToString());
|
|
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo004", iter->key().ToString());
|
|
ASSERT_EQ("v__4", iter->value().ToString());
|
|
|
|
iter->Seek("3000000000000bar");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("3000000000000bar", iter->key().ToString());
|
|
ASSERT_EQ("bar_v", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo005");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo005", iter->key().ToString());
|
|
ASSERT_EQ("v__5", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo006");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo005", iter->key().ToString());
|
|
ASSERT_EQ("v__5", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo008");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("1000000000foo008", iter->key().ToString());
|
|
ASSERT_EQ("v__8", iter->value().ToString());
|
|
|
|
iter->Seek("1000000000foo000");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("3000000000000bar", iter->key().ToString());
|
|
|
|
delete iter;
|
|
}
|
|
|
|
TEST(PlainTableDBTest, HashBucketConflict) {
|
|
for (size_t huge_page_tlb_size = 0; huge_page_tlb_size <= 2 * 1024 * 1024;
|
|
huge_page_tlb_size += 2 * 1024 * 1024) {
|
|
for (unsigned char i = 1; i <= 3; i++) {
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 16;
|
|
plain_table_options.bloom_bits_per_key = 0;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
plain_table_options.index_sparseness = 2 ^ i;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory(plain_table_options));
|
|
|
|
DestroyAndReopen(&options);
|
|
ASSERT_OK(Put("5000000000000fo0", "v1"));
|
|
ASSERT_OK(Put("5000000000000fo1", "v2"));
|
|
ASSERT_OK(Put("5000000000000fo2", "v"));
|
|
ASSERT_OK(Put("2000000000000fo0", "v3"));
|
|
ASSERT_OK(Put("2000000000000fo1", "v4"));
|
|
ASSERT_OK(Put("2000000000000fo2", "v"));
|
|
ASSERT_OK(Put("2000000000000fo3", "v"));
|
|
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
ASSERT_EQ("v1", Get("5000000000000fo0"));
|
|
ASSERT_EQ("v2", Get("5000000000000fo1"));
|
|
ASSERT_EQ("v3", Get("2000000000000fo0"));
|
|
ASSERT_EQ("v4", Get("2000000000000fo1"));
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get("5000000000000bar"));
|
|
ASSERT_EQ("NOT_FOUND", Get("2000000000000bar"));
|
|
ASSERT_EQ("NOT_FOUND", Get("5000000000000fo8"));
|
|
ASSERT_EQ("NOT_FOUND", Get("2000000000000fo8"));
|
|
|
|
ReadOptions ro;
|
|
Iterator* iter = dbfull()->NewIterator(ro);
|
|
|
|
iter->Seek("5000000000000fo0");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo0", iter->key().ToString());
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo1", iter->key().ToString());
|
|
|
|
iter->Seek("5000000000000fo1");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo1", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000fo0");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo0", iter->key().ToString());
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo1", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000fo1");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo1", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000bar");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo0", iter->key().ToString());
|
|
|
|
iter->Seek("5000000000000bar");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo0", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000fo8");
|
|
ASSERT_TRUE(!iter->Valid() ||
|
|
options.comparator->Compare(iter->key(), "20000001") > 0);
|
|
|
|
iter->Seek("5000000000000fo8");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("1000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("3000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("8000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
delete iter;
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(PlainTableDBTest, HashBucketConflictReverseSuffixComparator) {
|
|
for (size_t huge_page_tlb_size = 0; huge_page_tlb_size <= 2 * 1024 * 1024;
|
|
huge_page_tlb_size += 2 * 1024 * 1024) {
|
|
for (unsigned char i = 1; i <= 3; i++) {
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
SimpleSuffixReverseComparator comp;
|
|
options.comparator = ∁
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 16;
|
|
plain_table_options.bloom_bits_per_key = 0;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
plain_table_options.index_sparseness = 2 ^ i;
|
|
plain_table_options.huge_page_tlb_size = huge_page_tlb_size;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory(plain_table_options));
|
|
DestroyAndReopen(&options);
|
|
ASSERT_OK(Put("5000000000000fo0", "v1"));
|
|
ASSERT_OK(Put("5000000000000fo1", "v2"));
|
|
ASSERT_OK(Put("5000000000000fo2", "v"));
|
|
ASSERT_OK(Put("2000000000000fo0", "v3"));
|
|
ASSERT_OK(Put("2000000000000fo1", "v4"));
|
|
ASSERT_OK(Put("2000000000000fo2", "v"));
|
|
ASSERT_OK(Put("2000000000000fo3", "v"));
|
|
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
ASSERT_EQ("v1", Get("5000000000000fo0"));
|
|
ASSERT_EQ("v2", Get("5000000000000fo1"));
|
|
ASSERT_EQ("v3", Get("2000000000000fo0"));
|
|
ASSERT_EQ("v4", Get("2000000000000fo1"));
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get("5000000000000bar"));
|
|
ASSERT_EQ("NOT_FOUND", Get("2000000000000bar"));
|
|
ASSERT_EQ("NOT_FOUND", Get("5000000000000fo8"));
|
|
ASSERT_EQ("NOT_FOUND", Get("2000000000000fo8"));
|
|
|
|
ReadOptions ro;
|
|
Iterator* iter = dbfull()->NewIterator(ro);
|
|
|
|
iter->Seek("5000000000000fo1");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo1", iter->key().ToString());
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo0", iter->key().ToString());
|
|
|
|
iter->Seek("5000000000000fo1");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo1", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000fo1");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo1", iter->key().ToString());
|
|
iter->Next();
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo0", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000fo1");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo1", iter->key().ToString());
|
|
|
|
iter->Seek("2000000000000var");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("2000000000000fo3", iter->key().ToString());
|
|
|
|
iter->Seek("5000000000000var");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo2", iter->key().ToString());
|
|
|
|
std::string seek_key = "2000000000000bar";
|
|
iter->Seek(seek_key);
|
|
ASSERT_TRUE(!iter->Valid() ||
|
|
options.prefix_extractor->Transform(iter->key()) !=
|
|
options.prefix_extractor->Transform(seek_key));
|
|
|
|
iter->Seek("1000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("3000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("8000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
delete iter;
|
|
}
|
|
}
|
|
}
|
|
|
|
TEST(PlainTableDBTest, NonExistingKeyToNonEmptyBucket) {
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
// Set only one bucket to force bucket conflict.
|
|
// Test index interval for the same prefix to be 1, 2 and 4
|
|
PlainTableOptions plain_table_options;
|
|
plain_table_options.user_key_len = 16;
|
|
plain_table_options.bloom_bits_per_key = 0;
|
|
plain_table_options.hash_table_ratio = 0;
|
|
plain_table_options.index_sparseness = 5;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory(plain_table_options));
|
|
DestroyAndReopen(&options);
|
|
ASSERT_OK(Put("5000000000000fo0", "v1"));
|
|
ASSERT_OK(Put("5000000000000fo1", "v2"));
|
|
ASSERT_OK(Put("5000000000000fo2", "v3"));
|
|
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
ASSERT_EQ("v1", Get("5000000000000fo0"));
|
|
ASSERT_EQ("v2", Get("5000000000000fo1"));
|
|
ASSERT_EQ("v3", Get("5000000000000fo2"));
|
|
|
|
ASSERT_EQ("NOT_FOUND", Get("8000000000000bar"));
|
|
ASSERT_EQ("NOT_FOUND", Get("1000000000000bar"));
|
|
|
|
Iterator* iter = dbfull()->NewIterator(ReadOptions());
|
|
|
|
iter->Seek("5000000000000bar");
|
|
ASSERT_TRUE(iter->Valid());
|
|
ASSERT_EQ("5000000000000fo0", iter->key().ToString());
|
|
|
|
iter->Seek("5000000000000fo8");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("1000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
iter->Seek("8000000000000fo2");
|
|
ASSERT_TRUE(!iter->Valid());
|
|
|
|
delete iter;
|
|
}
|
|
|
|
static std::string Key(int i) {
|
|
char buf[100];
|
|
snprintf(buf, sizeof(buf), "key_______%06d", i);
|
|
return std::string(buf);
|
|
}
|
|
|
|
static std::string RandomString(Random* rnd, int len) {
|
|
std::string r;
|
|
test::RandomString(rnd, len, &r);
|
|
return r;
|
|
}
|
|
|
|
TEST(PlainTableDBTest, CompactionTrigger) {
|
|
Options options = CurrentOptions();
|
|
options.write_buffer_size = 100 << 10; //100KB
|
|
options.num_levels = 3;
|
|
options.max_mem_compaction_level = 0;
|
|
options.level0_file_num_compaction_trigger = 3;
|
|
Reopen(&options);
|
|
|
|
Random rnd(301);
|
|
|
|
for (int num = 0; num < options.level0_file_num_compaction_trigger - 1;
|
|
num++) {
|
|
std::vector<std::string> values;
|
|
// Write 120KB (12 values, each 10K)
|
|
for (int i = 0; i < 12; i++) {
|
|
values.push_back(RandomString(&rnd, 10000));
|
|
ASSERT_OK(Put(Key(i), values[i]));
|
|
}
|
|
dbfull()->TEST_WaitForFlushMemTable();
|
|
ASSERT_EQ(NumTableFilesAtLevel(0), num + 1);
|
|
}
|
|
|
|
//generate one more file in level-0, and should trigger level-0 compaction
|
|
std::vector<std::string> values;
|
|
for (int i = 0; i < 12; i++) {
|
|
values.push_back(RandomString(&rnd, 10000));
|
|
ASSERT_OK(Put(Key(i), values[i]));
|
|
}
|
|
dbfull()->TEST_WaitForCompact();
|
|
|
|
ASSERT_EQ(NumTableFilesAtLevel(0), 0);
|
|
ASSERT_EQ(NumTableFilesAtLevel(1), 1);
|
|
}
|
|
|
|
TEST(PlainTableDBTest, AdaptiveTable) {
|
|
Options options = CurrentOptions();
|
|
options.create_if_missing = true;
|
|
|
|
options.table_factory.reset(NewPlainTableFactory());
|
|
DestroyAndReopen(&options);
|
|
|
|
ASSERT_OK(Put("1000000000000foo", "v1"));
|
|
ASSERT_OK(Put("0000000000000bar", "v2"));
|
|
ASSERT_OK(Put("1000000000000foo", "v3"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
|
|
options.create_if_missing = false;
|
|
std::shared_ptr<TableFactory> dummy_factory;
|
|
std::shared_ptr<TableFactory> block_based_factory(
|
|
NewBlockBasedTableFactory());
|
|
options.table_factory.reset(NewAdaptiveTableFactory(
|
|
block_based_factory, dummy_factory, dummy_factory));
|
|
Reopen(&options);
|
|
ASSERT_EQ("v3", Get("1000000000000foo"));
|
|
ASSERT_EQ("v2", Get("0000000000000bar"));
|
|
|
|
ASSERT_OK(Put("2000000000000foo", "v4"));
|
|
ASSERT_OK(Put("3000000000000bar", "v5"));
|
|
dbfull()->TEST_FlushMemTable();
|
|
ASSERT_EQ("v4", Get("2000000000000foo"));
|
|
ASSERT_EQ("v5", Get("3000000000000bar"));
|
|
|
|
Reopen(&options);
|
|
ASSERT_EQ("v3", Get("1000000000000foo"));
|
|
ASSERT_EQ("v2", Get("0000000000000bar"));
|
|
ASSERT_EQ("v4", Get("2000000000000foo"));
|
|
ASSERT_EQ("v5", Get("3000000000000bar"));
|
|
|
|
options.table_factory.reset(NewBlockBasedTableFactory());
|
|
Reopen(&options);
|
|
ASSERT_NE("v3", Get("1000000000000foo"));
|
|
|
|
options.table_factory.reset(NewPlainTableFactory());
|
|
Reopen(&options);
|
|
ASSERT_NE("v5", Get("3000000000000bar"));
|
|
}
|
|
|
|
} // namespace rocksdb
|
|
|
|
int main(int argc, char** argv) {
|
|
return rocksdb::test::RunAllTests();
|
|
}
|