$Cases = @{}
if($Suites.Count -gt 0) {
- # Will teh first one
+ # Will the first one
ForEach($e in $Suites.Keys) {
$exe_name = $e
$Cases = $Suites[$e]
// If DB just falled into the stop condition, we need to further reduce
// the write rate to avoid the stop condition.
if (penalize_stop) {
- // Penalize the near stop or stop condition by more agressive slowdown.
+ // Penalize the near stop or stop condition by more aggressive slowdown.
// This is to provide the long term slowdown increase signal.
// The penalty is more than the reward of recovering to the normal
// condition.
const std::string& cf_name, const MutableCFOptions& mutable_cf_options,
VersionStorageInfo* vstorage, double score,
const std::vector<SortedRun>& sorted_runs, LogBuffer* log_buffer) {
- // percentage flexibilty while reducing size amplification
+ // percentage flexibility while reducing size amplification
uint64_t ratio =
ioptions_.compaction_options_universal.max_size_amplification_percent;
// A global method that can dump out the build version
void DumpRocksDBBuildVersion(Logger * log) {
#if !defined(IOS_CROSS_COMPILE)
- // if we compile with Xcode, we don't run build_detect_vesion, so we don't
+ // if we compile with Xcode, we don't run build_detect_version, so we don't
// generate util/build_version.cc
ROCKS_LOG_HEADER(log, "RocksDB version: %d.%d.%d\n", ROCKSDB_MAJOR,
ROCKSDB_MINOR, ROCKSDB_PATCH);
// REQUIRES: Mutex held
void UpdateStats();
- // Cleanup after successfull/failed job
+ // Cleanup after successful/failed job
void Cleanup(const Status& status);
VersionEdit* edit() { return &edit_; }
// If enable_pipelined_write is true, separate write thread queue is
// maintained for WAL write and memtable write. A write thread first enter WAL
// writer queue and then memtable writer queue. Pending thread on the WAL
- // writer queue thus only have to wait for previous writers to finish thier
+ // writer queue thus only have to wait for previous writers to finish their
// WAL writing but not the memtable writing. Enabling the feature may improve
// write throughput and reduce latency of the prepare phase of two-phase
// commit.
jobject jbackup_info_handle_list =
env->NewObject(jarray_list_clazz, cstr_mid, backup_infos.size());
if(env->ExceptionCheck()) {
- // exception occured constructing object
+ // exception occurred constructing object
return nullptr;
}
backup_info.size,
backup_info.number_files);
if(env->ExceptionCheck()) {
- // exception occured constructing object
+ // exception occurred constructing object
if(obj != nullptr) {
env->DeleteLocalRef(obj);
}
jboolean rs =
env->CallBooleanMethod(jbackup_info_handle_list, add_mid, obj);
if(env->ExceptionCheck() || rs == JNI_FALSE) {
- // exception occured calling method, or could not add
+ // exception occurred calling method, or could not add
if(obj != nullptr) {
env->DeleteLocalRef(obj);
}
env->GetStaticFieldID(jclazz, name,
"Lorg/rocksdb/WBWIRocksIterator$WriteType;");
if(env->ExceptionCheck()) {
- // exception occured while getting field
+ // exception occurred while getting field
return nullptr;
} else if(jfid == nullptr) {
return nullptr;
jfieldID jfid =
env->GetStaticFieldID(jclazz, name, "Lorg/rocksdb/InfoLogLevel;");
if(env->ExceptionCheck()) {
- // exception occured while getting field
+ // exception occurred while getting field
return nullptr;
} else if(jfid == nullptr) {
return nullptr;
env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT);
if(has_exception == JNI_TRUE) {
- // exception occured
+ // exception occurred
env->ReleaseStringUTFChars(jdb_path, db_path);
return nullptr;
}
rocksdb::JniUtil::copyStrings(env, jfile_path_list, jfile_path_list_len,
&has_exception);
if(has_exception == JNI_TRUE) {
- // exception occured
+ // exception occurred
return;
}
env->ReleaseLongArrayElements(jcolumn_options, jco, JNI_ABORT);
if(has_exception == JNI_TRUE) {
- // exception occured
+ // exception occurred
env->ReleaseStringUTFChars(jdb_path, db_path);
return nullptr;
}
}
/**
- * Percentage flexibilty while comparing file size. If the candidate file(s)
+ * Percentage flexibility while comparing file size. If the candidate file(s)
* size is 1% smaller than the next file's size, then include next file into
* this candidate set.
*
}
/**
- * Percentage flexibilty while comparing file size. If the candidate file(s)
+ * Percentage flexibility while comparing file size. If the candidate file(s)
* size is 1% smaller than the next file's size, then include next file into
* this candidate set.
*
ASSERT_EQ(new_cf_opt.arena_block_size, 21 * tera);
// Nested block based table options
- // Emtpy
+ // Empty
ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt,
"write_buffer_size=10;max_write_buffer_number=16;"
"block_based_table_factory={};arena_block_size=1024",
ASSERT_OK(RocksDBOptionsParser::VerifyCFOptions(base_cf_opt, new_cf_opt));
// Nested plain table options
- // Emtpy
+ // Empty
ASSERT_OK(GetColumnFamilyOptionsFromString(base_cf_opt,
"write_buffer_size=10;max_write_buffer_number=16;"
"plain_table_factory={};arena_block_size=1024",
### CreateColumnFamily
- # Create a new column familiy for the database.
+ # Create a new column family for the database.
#
# column_family_name (string) - Name of the new column family.
#
### Installation
NOTE: If your default `python` binary is not a version of python2, add
-the arguments `--python /path/to/python2` to the the `node-gyp` commands.
+the arguments `--python /path/to/python2` to the `node-gyp` commands.
1. Make sure you have the static library (i.e. "librocksdb.a") in the root
directory of your rocksdb installation. If not, `cd` there and run
//
#pragma once
#if !defined(IOS_CROSS_COMPILE)
-// if we compile with Xcode, we don't run build_detect_vesion, so we don't
+// if we compile with Xcode, we don't run build_detect_version, so we don't
// generate these variables
// this variable tells us about the git revision
extern const char* rocksdb_build_git_sha;
// By default, xxHash library provides endian-independent Hash values, based on little-endian convention.
// Results are therefore identical for little-endian and big-endian CPU.
// This comes at a performance cost for big-endian CPU, since some swapping is required to emulate little-endian format.
-// Should endian-independance be of no importance for your application, you may set the #define below to 1.
+// Should endian-independence be of no importance for your application, you may set the #define below to 1.
// It will improve speed for Big-endian CPU.
// This option has no impact on Little_Endian CPU.
#define XXH_FORCE_NATIVE_FORMAT 0
// IMPORTANT NOTE: Secondary index column families should be very small and
// generally fit in memory. Assume that accessing secondary index column
// families is much faster than accessing primary index (data heap) column
-// family. Accessing a key (i.e. checking for existance) from a column family in
+// family. Accessing a key (i.e. checking for existence) from a column family in
// RocksDB is not much faster than accessing both key and value since they are
// kept together and loaded from storage together.