blob: a1eb7b8543af23aac23ef0d6250ca0d803c7cab3 [file] [log] [blame]
/*
* Copyright (c) 2014, 2024, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*
*/
#include "precompiled.hpp"
#include "code/codeCache.hpp"
#include "code/nmethod.hpp"
#include "gc/g1/g1CodeRootSet.hpp"
#include "gc/g1/heapRegion.hpp"
#include "memory/allocation.hpp"
#include "oops/oop.inline.hpp"
#include "runtime/atomic.hpp"
#include "utilities/concurrentHashTable.inline.hpp"
#include "utilities/concurrentHashTableTasks.inline.hpp"
class G1CodeRootSetHashTableConfig : public StackObj {
public:
using Value = nmethod*;
static uintx get_hash(Value const& value, bool* is_dead);
static void* allocate_node(void* context, size_t size, Value const& value) {
return AllocateHeap(size, mtGC);
}
static void free_node(void* context, void* memory, Value const& value) {
FreeHeap(memory);
}
};
// Storage container for the code root set.
class G1CodeRootSetHashTable : public CHeapObj<mtGC> {
using HashTable = ConcurrentHashTable<G1CodeRootSetHashTableConfig, mtGC>;
using HashTableScanTask = HashTable::ScanTask;
// Default (log2) number of buckets; small since typically we do not expect many
// entries.
static const size_t Log2DefaultNumBuckets = 2;
static const uint BucketClaimSize = 16;
HashTable _table;
HashTableScanTask _table_scanner;
size_t volatile _num_entries;
bool is_empty() const { return number_of_entries() == 0; }
class HashTableLookUp : public StackObj {
nmethod* _nmethod;
public:
explicit HashTableLookUp(nmethod* nmethod) : _nmethod(nmethod) { }
uintx get_hash() const;
bool equals(nmethod** value);
bool is_dead(nmethod** value) const { return false; }
};
class HashTableIgnore : public StackObj {
public:
HashTableIgnore() { }
void operator()(nmethod** value) { /* do nothing */ }
};
public:
G1CodeRootSetHashTable() :
_table(Mutex::service-1,
nullptr,
Log2DefaultNumBuckets,
false /* enable_statistics */),
_table_scanner(&_table, BucketClaimSize), _num_entries(0) {
clear();
}
// Robert Jenkins 1996 & Thomas Wang 1997
// http://web.archive.org/web/20071223173210/http://www.concentric.net/~Ttwang/tech/inthash.htm
static uint32_t hash(uint32_t key) {
key = ~key + (key << 15);
key = key ^ (key >> 12);
key = key + (key << 2);
key = key ^ (key >> 4);
key = key * 2057;
key = key ^ (key >> 16);
return key;
}
static uintx get_hash(nmethod* nmethod) {
uintptr_t value = (uintptr_t)nmethod;
// The CHT only uses the bits smaller than HashTable::DEFAULT_MAX_SIZE_LOG2, so
// try to increase the randomness by incorporating the upper bits of the
// address too.
STATIC_ASSERT(HashTable::DEFAULT_MAX_SIZE_LOG2 <= sizeof(uint32_t) * BitsPerByte);
#ifdef _LP64
return hash((uint32_t)value ^ (uint32_t(value >> 32)));
#else
return hash((uint32_t)value);
#endif
}
void insert(nmethod* method) {
HashTableLookUp lookup(method);
bool grow_hint = false;
bool inserted = _table.insert(Thread::current(), lookup, method, &grow_hint);
if (inserted) {
Atomic::inc(&_num_entries);
}
if (grow_hint) {
_table.grow(Thread::current());
}
}
bool remove(nmethod* method) {
HashTableLookUp lookup(method);
bool removed = _table.remove(Thread::current(), lookup);
if (removed) {
Atomic::dec(&_num_entries);
}
return removed;
}
bool contains(nmethod* method) {
HashTableLookUp lookup(method);
HashTableIgnore ignore;
return _table.get(Thread::current(), lookup, ignore);
}
void clear() {
// Remove all entries.
auto always_true = [] (nmethod** value) {
return true;
};
clean(always_true);
}
void iterate_at_safepoint(CodeBlobClosure* blk) {
assert_at_safepoint();
// A lot of code root sets are typically empty.
if (is_empty()) {
return;
}
auto do_value =
[&] (nmethod** value) {
blk->do_code_blob(*value);
return true;
};
_table_scanner.do_safepoint_scan(do_value);
}
// Removes entries as indicated by the given EVAL closure.
template <class EVAL>
void clean(EVAL& eval) {
// A lot of code root sets are typically empty.
if (is_empty()) {
return;
}
size_t num_deleted = 0;
auto do_delete =
[&] (nmethod** value) {
num_deleted++;
};
bool succeeded = _table.try_bulk_delete(Thread::current(), eval, do_delete);
guarantee(succeeded, "unable to clean table");
if (num_deleted != 0) {
size_t current_size = Atomic::sub(&_num_entries, num_deleted);
shrink_to_match(current_size);
}
}
// Removes dead/unlinked entries.
void bulk_remove() {
auto delete_check = [&] (nmethod** value) {
return (*value)->is_unlinked();
};
clean(delete_check);
}
// Calculate the log2 of the table size we want to shrink to.
size_t log2_target_shrink_size(size_t current_size) const {
// A table with the new size should be at most filled by this factor. Otherwise
// we would grow again quickly.
const float WantedLoadFactor = 0.5;
size_t min_expected_size = checked_cast<size_t>(ceil(current_size / WantedLoadFactor));
size_t result = Log2DefaultNumBuckets;
if (min_expected_size != 0) {
size_t log2_bound = checked_cast<size_t>(log2i_exact(round_up_power_of_2(min_expected_size)));
result = clamp(log2_bound, Log2DefaultNumBuckets, HashTable::DEFAULT_MAX_SIZE_LOG2);
}
return result;
}
// Shrink to keep table size appropriate to the given number of entries.
void shrink_to_match(size_t current_size) {
size_t prev_log2size = _table.get_size_log2(Thread::current());
size_t new_log2_table_size = log2_target_shrink_size(current_size);
if (new_log2_table_size < prev_log2size) {
_table.shrink(Thread::current(), new_log2_table_size);
}
}
void reset_table_scanner() {
_table_scanner.set(&_table, BucketClaimSize);
}
size_t mem_size() { return sizeof(*this) + _table.get_mem_size(Thread::current()); }
size_t number_of_entries() const { return Atomic::load(&_num_entries); }
};
uintx G1CodeRootSetHashTable::HashTableLookUp::get_hash() const {
return G1CodeRootSetHashTable::get_hash(_nmethod);
}
bool G1CodeRootSetHashTable::HashTableLookUp::equals(nmethod** value) {
return *value == _nmethod;
}
uintx G1CodeRootSetHashTableConfig::get_hash(Value const& value, bool* is_dead) {
*is_dead = false;
return G1CodeRootSetHashTable::get_hash(value);
}
size_t G1CodeRootSet::length() const { return _table->number_of_entries(); }
void G1CodeRootSet::add(nmethod* method) {
if (!contains(method)) {
assert(!_is_iterating, "must be");
_table->insert(method);
}
}
G1CodeRootSet::G1CodeRootSet() :
_table(new G1CodeRootSetHashTable())
DEBUG_ONLY(COMMA _is_iterating(false)) { }
G1CodeRootSet::~G1CodeRootSet() {
delete _table;
}
bool G1CodeRootSet::remove(nmethod* method) {
assert(!_is_iterating, "should not mutate while iterating the table");
return _table->remove(method);
}
void G1CodeRootSet::bulk_remove() {
assert(!_is_iterating, "should not mutate while iterating the table");
_table->bulk_remove();
}
bool G1CodeRootSet::contains(nmethod* method) {
return _table->contains(method);
}
void G1CodeRootSet::clear() {
assert(!_is_iterating, "should not mutate while iterating the table");
_table->clear();
}
size_t G1CodeRootSet::mem_size() {
return sizeof(*this) + _table->mem_size();
}
void G1CodeRootSet::reset_table_scanner() {
_table->reset_table_scanner();
}
void G1CodeRootSet::nmethods_do(CodeBlobClosure* blk) const {
DEBUG_ONLY(_is_iterating = true;)
_table->iterate_at_safepoint(blk);
DEBUG_ONLY(_is_iterating = false;)
}
class CleanCallback : public StackObj {
NONCOPYABLE(CleanCallback); // can not copy, _blobs will point to old copy
class PointsIntoHRDetectionClosure : public OopClosure {
HeapRegion* _hr;
template <typename T>
void do_oop_work(T* p) {
if (_hr->is_in(RawAccess<>::oop_load(p))) {
_points_into = true;
}
}
public:
bool _points_into;
PointsIntoHRDetectionClosure(HeapRegion* hr) : _hr(hr), _points_into(false) {}
void do_oop(narrowOop* o) { do_oop_work(o); }
void do_oop(oop* o) { do_oop_work(o); }
};
PointsIntoHRDetectionClosure _detector;
CodeBlobToOopClosure _blobs;
public:
CleanCallback(HeapRegion* hr) : _detector(hr), _blobs(&_detector, !CodeBlobToOopClosure::FixRelocations) {}
bool operator()(nmethod** value) {
_detector._points_into = false;
_blobs.do_code_blob(*value);
return !_detector._points_into;
}
};
void G1CodeRootSet::clean(HeapRegion* owner) {
assert(!_is_iterating, "should not mutate while iterating the table");
CleanCallback eval(owner);
_table->clean(eval);
}