rocksdb/utilities/transactions/lock/range/range_tree/lib/locktree/treenode.cc
Sergei Petrunia 1022090981 Apply the changes from: PS-5501 : Re-license PerconaFT 'locktree' to Apache V2 (#7801)
Summary:
commit d5178f513c0b4144a5ac9358ec0f6a3b54a28e76
Author: George O. Lorch III <george.lorch@percona.com>
Date:   Tue Mar 19 12:18:40 2019 -0700

    PS-5501 : Re-license PerconaFT 'locktree' to Apache V2

    - Fixed some incomplete relicensed files from previous round.

    - Added missing license text to some.

    - Relicensed more files to Apache V2 that locktree depends on.

Pull Request resolved: https://github.com/facebook/rocksdb/pull/7801

Reviewed By: jay-zhuang

Differential Revision: D25682430

Pulled By: cheng-chang

fbshipit-source-id: deb8a0de3e76f3638672997bfbd300e2fffbe5f5
2020-12-22 14:47:41 -08:00

521 lines
16 KiB
C++

/* -*- mode: C++; c-basic-offset: 4; indent-tabs-mode: nil -*- */
// vim: ft=cpp:expandtab:ts=8:sw=4:softtabstop=4:
#ifndef ROCKSDB_LITE
#ifndef OS_WIN
#ident "$Id$"
/*======
This file is part of PerconaFT.
Copyright (c) 2006, 2015, Percona and/or its affiliates. All rights reserved.
PerconaFT is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License, version 2,
as published by the Free Software Foundation.
PerconaFT is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with PerconaFT. If not, see <http://www.gnu.org/licenses/>.
----------------------------------------
PerconaFT is free software: you can redistribute it and/or modify
it under the terms of the GNU Affero General Public License, version 3,
as published by the Free Software Foundation.
PerconaFT is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU Affero General Public License for more details.
You should have received a copy of the GNU Affero General Public License
along with PerconaFT. If not, see <http://www.gnu.org/licenses/>.
----------------------------------------
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
======= */
#ident \
"Copyright (c) 2006, 2015, Percona and/or its affiliates. All rights reserved."
#include "treenode.h"
#include "../portability/toku_race_tools.h"
namespace toku {
// TODO: source location info might have to be pulled up one caller
// to be useful
void treenode::mutex_lock(void) { toku_mutex_lock(&m_mutex); }
void treenode::mutex_unlock(void) { toku_mutex_unlock(&m_mutex); }
void treenode::init(const comparator *cmp) {
m_txnid = TXNID_NONE;
m_is_root = false;
m_is_empty = true;
m_cmp = cmp;
m_is_shared = false;
m_owners = nullptr;
// use an adaptive mutex at each node since we expect the time the
// lock is held to be relatively short compared to a context switch.
// indeed, this improves performance at high thread counts considerably.
memset(&m_mutex, 0, sizeof(toku_mutex_t));
toku_pthread_mutexattr_t attr;
toku_mutexattr_init(&attr);
toku_mutexattr_settype(&attr, TOKU_MUTEX_ADAPTIVE);
toku_mutex_init(treenode_mutex_key, &m_mutex, &attr);
toku_mutexattr_destroy(&attr);
m_left_child.set(nullptr);
m_right_child.set(nullptr);
}
void treenode::create_root(const comparator *cmp) {
init(cmp);
m_is_root = true;
}
void treenode::destroy_root(void) {
invariant(is_root());
invariant(is_empty());
toku_mutex_destroy(&m_mutex);
m_cmp = nullptr;
}
void treenode::set_range_and_txnid(const keyrange &range, TXNID txnid,
bool is_shared) {
// allocates a new copy of the range for this node
m_range.create_copy(range);
m_txnid = txnid;
m_is_shared = is_shared;
m_is_empty = false;
}
bool treenode::is_root(void) { return m_is_root; }
bool treenode::is_empty(void) { return m_is_empty; }
bool treenode::range_overlaps(const keyrange &range) {
return m_range.overlaps(*m_cmp, range);
}
treenode *treenode::alloc(const comparator *cmp, const keyrange &range,
TXNID txnid, bool is_shared) {
treenode *XCALLOC(node);
node->init(cmp);
node->set_range_and_txnid(range, txnid, is_shared);
return node;
}
void treenode::swap_in_place(treenode *node1, treenode *node2) {
keyrange tmp_range = node1->m_range;
TXNID tmp_txnid = node1->m_txnid;
node1->m_range = node2->m_range;
node1->m_txnid = node2->m_txnid;
node2->m_range = tmp_range;
node2->m_txnid = tmp_txnid;
bool tmp_is_shared = node1->m_is_shared;
node1->m_is_shared = node2->m_is_shared;
node2->m_is_shared = tmp_is_shared;
auto tmp_m_owners = node1->m_owners;
node1->m_owners = node2->m_owners;
node2->m_owners = tmp_m_owners;
}
bool treenode::add_shared_owner(TXNID txnid) {
assert(m_is_shared);
if (txnid == m_txnid)
return false; // acquiring a lock on the same range by the same trx
if (m_txnid != TXNID_SHARED) {
m_owners = new TxnidVector;
m_owners->insert(m_txnid);
m_txnid = TXNID_SHARED;
}
m_owners->insert(txnid);
return true;
}
void treenode::free(treenode *node) {
// destroy the range, freeing any copied keys
node->m_range.destroy();
if (node->m_owners) {
delete node->m_owners;
node->m_owners = nullptr; // need this?
}
// the root is simply marked as empty.
if (node->is_root()) {
// PORT toku_mutex_assert_locked(&node->m_mutex);
node->m_is_empty = true;
} else {
// PORT toku_mutex_assert_unlocked(&node->m_mutex);
toku_mutex_destroy(&node->m_mutex);
toku_free(node);
}
}
uint32_t treenode::get_depth_estimate(void) const {
const uint32_t left_est = m_left_child.depth_est;
const uint32_t right_est = m_right_child.depth_est;
return (left_est > right_est ? left_est : right_est) + 1;
}
treenode *treenode::find_node_with_overlapping_child(
const keyrange &range, const keyrange::comparison *cmp_hint) {
// determine which child to look at based on a comparison. if we were
// given a comparison hint, use that. otherwise, compare them now.
keyrange::comparison c =
cmp_hint ? *cmp_hint : range.compare(*m_cmp, m_range);
treenode *child;
if (c == keyrange::comparison::LESS_THAN) {
child = lock_and_rebalance_left();
} else {
// The caller (locked_keyrange::acquire) handles the case where
// the root of the locked_keyrange is the node that overlaps.
// range is guaranteed not to overlap this node.
invariant(c == keyrange::comparison::GREATER_THAN);
child = lock_and_rebalance_right();
}
// if the search would lead us to an empty subtree (child == nullptr),
// or the child overlaps, then we know this node is the parent we want.
// otherwise we need to recur into that child.
if (child == nullptr) {
return this;
} else {
c = range.compare(*m_cmp, child->m_range);
if (c == keyrange::comparison::EQUALS ||
c == keyrange::comparison::OVERLAPS) {
child->mutex_unlock();
return this;
} else {
// unlock this node before recurring into the locked child,
// passing in a comparison hint since we just comapred range
// to the child's range.
mutex_unlock();
return child->find_node_with_overlapping_child(range, &c);
}
}
}
bool treenode::insert(const keyrange &range, TXNID txnid, bool is_shared) {
int rc = true;
// choose a child to check. if that child is null, then insert the new node
// there. otherwise recur down that child's subtree
keyrange::comparison c = range.compare(*m_cmp, m_range);
if (c == keyrange::comparison::LESS_THAN) {
treenode *left_child = lock_and_rebalance_left();
if (left_child == nullptr) {
left_child = treenode::alloc(m_cmp, range, txnid, is_shared);
m_left_child.set(left_child);
} else {
left_child->insert(range, txnid, is_shared);
left_child->mutex_unlock();
}
} else if (c == keyrange::comparison::GREATER_THAN) {
// invariant(c == keyrange::comparison::GREATER_THAN);
treenode *right_child = lock_and_rebalance_right();
if (right_child == nullptr) {
right_child = treenode::alloc(m_cmp, range, txnid, is_shared);
m_right_child.set(right_child);
} else {
right_child->insert(range, txnid, is_shared);
right_child->mutex_unlock();
}
} else if (c == keyrange::comparison::EQUALS) {
invariant(is_shared);
invariant(m_is_shared);
rc = add_shared_owner(txnid);
} else {
invariant(0);
}
return rc;
}
treenode *treenode::find_child_at_extreme(int direction, treenode **parent) {
treenode *child =
direction > 0 ? m_right_child.get_locked() : m_left_child.get_locked();
if (child) {
*parent = this;
treenode *child_extreme = child->find_child_at_extreme(direction, parent);
child->mutex_unlock();
return child_extreme;
} else {
return this;
}
}
treenode *treenode::find_leftmost_child(treenode **parent) {
return find_child_at_extreme(-1, parent);
}
treenode *treenode::find_rightmost_child(treenode **parent) {
return find_child_at_extreme(1, parent);
}
treenode *treenode::remove_root_of_subtree() {
// if this node has no children, just free it and return null
if (m_left_child.ptr == nullptr && m_right_child.ptr == nullptr) {
// treenode::free requires that non-root nodes are unlocked
if (!is_root()) {
mutex_unlock();
}
treenode::free(this);
return nullptr;
}
// we have a child, so get either the in-order successor or
// predecessor of this node to be our replacement.
// replacement_parent is updated by the find functions as
// they recur down the tree, so initialize it to this.
treenode *child, *replacement;
treenode *replacement_parent = this;
if (m_left_child.ptr != nullptr) {
child = m_left_child.get_locked();
replacement = child->find_rightmost_child(&replacement_parent);
invariant(replacement == child || replacement_parent != this);
// detach the replacement from its parent
if (replacement_parent == this) {
m_left_child = replacement->m_left_child;
} else {
replacement_parent->m_right_child = replacement->m_left_child;
}
} else {
child = m_right_child.get_locked();
replacement = child->find_leftmost_child(&replacement_parent);
invariant(replacement == child || replacement_parent != this);
// detach the replacement from its parent
if (replacement_parent == this) {
m_right_child = replacement->m_right_child;
} else {
replacement_parent->m_left_child = replacement->m_right_child;
}
}
child->mutex_unlock();
// swap in place with the detached replacement, then destroy it
treenode::swap_in_place(replacement, this);
treenode::free(replacement);
return this;
}
void treenode::recursive_remove(void) {
treenode *left = m_left_child.ptr;
if (left) {
left->recursive_remove();
}
m_left_child.set(nullptr);
treenode *right = m_right_child.ptr;
if (right) {
right->recursive_remove();
}
m_right_child.set(nullptr);
// we do not take locks on the way down, so we know non-root nodes
// are unlocked here and the caller is required to pass a locked
// root, so this free is correct.
treenode::free(this);
}
void treenode::remove_shared_owner(TXNID txnid) {
assert(m_owners->size() > 1);
m_owners->erase(txnid);
assert(m_owners->size() > 0);
/* if there is just one owner left, move it to m_txnid */
if (m_owners->size() == 1) {
m_txnid = *m_owners->begin();
delete m_owners;
m_owners = nullptr;
}
}
treenode *treenode::remove(const keyrange &range, TXNID txnid) {
treenode *child;
// if the range is equal to this node's range, then just remove
// the root of this subtree. otherwise search down the tree
// in either the left or right children.
keyrange::comparison c = range.compare(*m_cmp, m_range);
switch (c) {
case keyrange::comparison::EQUALS: {
// if we are the only owners, remove. Otherwise, just remove
// us from the owners list.
if (txnid != TXNID_ANY && has_multiple_owners()) {
remove_shared_owner(txnid);
return this;
} else {
return remove_root_of_subtree();
}
}
case keyrange::comparison::LESS_THAN:
child = m_left_child.get_locked();
invariant_notnull(child);
child = child->remove(range, txnid);
// unlock the child if there still is one.
// regardless, set the right child pointer
if (child) {
child->mutex_unlock();
}
m_left_child.set(child);
break;
case keyrange::comparison::GREATER_THAN:
child = m_right_child.get_locked();
invariant_notnull(child);
child = child->remove(range, txnid);
// unlock the child if there still is one.
// regardless, set the right child pointer
if (child) {
child->mutex_unlock();
}
m_right_child.set(child);
break;
case keyrange::comparison::OVERLAPS:
// shouldn't be overlapping, since the tree is
// non-overlapping and this range must exist
abort();
}
return this;
}
bool treenode::left_imbalanced(int threshold) const {
uint32_t left_depth = m_left_child.depth_est;
uint32_t right_depth = m_right_child.depth_est;
return m_left_child.ptr != nullptr && left_depth > threshold + right_depth;
}
bool treenode::right_imbalanced(int threshold) const {
uint32_t left_depth = m_left_child.depth_est;
uint32_t right_depth = m_right_child.depth_est;
return m_right_child.ptr != nullptr && right_depth > threshold + left_depth;
}
// effect: rebalances the subtree rooted at this node
// using AVL style O(1) rotations. unlocks this
// node if it is not the new root of the subtree.
// requires: node is locked by this thread, children are not
// returns: locked root node of the rebalanced tree
treenode *treenode::maybe_rebalance(void) {
// if we end up not rotating at all, the new root is this
treenode *new_root = this;
treenode *child = nullptr;
if (left_imbalanced(IMBALANCE_THRESHOLD)) {
child = m_left_child.get_locked();
if (child->right_imbalanced(0)) {
treenode *grandchild = child->m_right_child.get_locked();
child->m_right_child = grandchild->m_left_child;
grandchild->m_left_child.set(child);
m_left_child = grandchild->m_right_child;
grandchild->m_right_child.set(this);
new_root = grandchild;
} else {
m_left_child = child->m_right_child;
child->m_right_child.set(this);
new_root = child;
}
} else if (right_imbalanced(IMBALANCE_THRESHOLD)) {
child = m_right_child.get_locked();
if (child->left_imbalanced(0)) {
treenode *grandchild = child->m_left_child.get_locked();
child->m_left_child = grandchild->m_right_child;
grandchild->m_right_child.set(child);
m_right_child = grandchild->m_left_child;
grandchild->m_left_child.set(this);
new_root = grandchild;
} else {
m_right_child = child->m_left_child;
child->m_left_child.set(this);
new_root = child;
}
}
// up to three nodes may be locked.
// - this
// - child
// - grandchild (but if it is locked, its the new root)
//
// one of them is the new root. we unlock everything except the new root.
if (child && child != new_root) {
TOKU_VALGRIND_RESET_MUTEX_ORDERING_INFO(&child->m_mutex);
child->mutex_unlock();
}
if (this != new_root) {
TOKU_VALGRIND_RESET_MUTEX_ORDERING_INFO(&m_mutex);
mutex_unlock();
}
TOKU_VALGRIND_RESET_MUTEX_ORDERING_INFO(&new_root->m_mutex);
return new_root;
}
treenode *treenode::lock_and_rebalance_left(void) {
treenode *child = m_left_child.get_locked();
if (child) {
treenode *new_root = child->maybe_rebalance();
m_left_child.set(new_root);
child = new_root;
}
return child;
}
treenode *treenode::lock_and_rebalance_right(void) {
treenode *child = m_right_child.get_locked();
if (child) {
treenode *new_root = child->maybe_rebalance();
m_right_child.set(new_root);
child = new_root;
}
return child;
}
void treenode::child_ptr::set(treenode *node) {
ptr = node;
depth_est = ptr ? ptr->get_depth_estimate() : 0;
}
treenode *treenode::child_ptr::get_locked(void) {
if (ptr) {
ptr->mutex_lock();
depth_est = ptr->get_depth_estimate();
}
return ptr;
}
} /* namespace toku */
#endif // OS_WIN
#endif // ROCKSDB_LITE