408 lines
14 KiB
C++
408 lines
14 KiB
C++
/* Copyright (C) 2017 The Android Open Source Project
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This file implements interfaces from the file jvmti.h. This implementation
|
|
* is licensed under the same terms as the file jvmti.h. The
|
|
* copyright and license information for the file jvmti.h follows.
|
|
*
|
|
* Copyright (c) 2003, 2011, Oracle and/or its affiliates. All rights reserved.
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
* published by the Free Software Foundation. Oracle designates this
|
|
* particular file as subject to the "Classpath" exception as provided
|
|
* by Oracle in the LICENSE file that accompanied this code.
|
|
*
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
* accompanied this code).
|
|
*
|
|
* You should have received a copy of the GNU General Public License version
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
* questions.
|
|
*/
|
|
|
|
#ifndef ART_OPENJDKJVMTI_JVMTI_WEAK_TABLE_INL_H_
|
|
#define ART_OPENJDKJVMTI_JVMTI_WEAK_TABLE_INL_H_
|
|
|
|
#include "jvmti_weak_table.h"
|
|
|
|
#include <limits>
|
|
|
|
#include <android-base/logging.h>
|
|
|
|
#include "art_jvmti.h"
|
|
#include "gc/allocation_listener.h"
|
|
#include "instrumentation.h"
|
|
#include "jni/jni_env_ext-inl.h"
|
|
#include "jvmti_allocator.h"
|
|
#include "mirror/class.h"
|
|
#include "mirror/object.h"
|
|
#include "nativehelper/scoped_local_ref.h"
|
|
#include "runtime.h"
|
|
|
|
namespace openjdkjvmti {
|
|
|
|
template <typename T>
|
|
void JvmtiWeakTable<T>::Lock() {
|
|
allow_disallow_lock_.ExclusiveLock(art::Thread::Current());
|
|
}
|
|
template <typename T>
|
|
void JvmtiWeakTable<T>::Unlock() {
|
|
allow_disallow_lock_.ExclusiveUnlock(art::Thread::Current());
|
|
}
|
|
template <typename T>
|
|
void JvmtiWeakTable<T>::AssertLocked() {
|
|
allow_disallow_lock_.AssertHeld(art::Thread::Current());
|
|
}
|
|
|
|
template <typename T>
|
|
void JvmtiWeakTable<T>::UpdateTableWithReadBarrier() {
|
|
update_since_last_sweep_ = true;
|
|
|
|
auto WithReadBarrierUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root,
|
|
art::mirror::Object* original_obj ATTRIBUTE_UNUSED)
|
|
REQUIRES_SHARED(art::Locks::mutator_lock_) {
|
|
return original_root.Read<art::kWithReadBarrier>();
|
|
};
|
|
|
|
UpdateTableWith<decltype(WithReadBarrierUpdater), kIgnoreNull>(WithReadBarrierUpdater);
|
|
}
|
|
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::GetTagSlowPath(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T* result) {
|
|
// Under concurrent GC, there is a window between moving objects and sweeping of system
|
|
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
|
|
// but still have from-space pointers in the table. Explicitly update the table once.
|
|
// Note: this will keep *all* objects in the table live, but should be a rare occurrence.
|
|
UpdateTableWithReadBarrier();
|
|
return GetTagLocked(self, obj, result);
|
|
}
|
|
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::Remove(art::ObjPtr<art::mirror::Object> obj, /* out */ T* tag) {
|
|
art::Thread* self = art::Thread::Current();
|
|
art::MutexLock mu(self, allow_disallow_lock_);
|
|
Wait(self);
|
|
|
|
return RemoveLocked(self, obj, tag);
|
|
}
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::RemoveLocked(art::ObjPtr<art::mirror::Object> obj, T* tag) {
|
|
art::Thread* self = art::Thread::Current();
|
|
allow_disallow_lock_.AssertHeld(self);
|
|
Wait(self);
|
|
|
|
return RemoveLocked(self, obj, tag);
|
|
}
|
|
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::RemoveLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T* tag) {
|
|
auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
|
|
if (it != tagged_objects_.end()) {
|
|
if (tag != nullptr) {
|
|
*tag = it->second;
|
|
}
|
|
tagged_objects_.erase(it);
|
|
return true;
|
|
}
|
|
|
|
if (art::kUseReadBarrier && self->GetIsGcMarking() && !update_since_last_sweep_) {
|
|
// Under concurrent GC, there is a window between moving objects and sweeping of system
|
|
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
|
|
// but still have from-space pointers in the table. Explicitly update the table once.
|
|
// Note: this will keep *all* objects in the table live, but should be a rare occurrence.
|
|
|
|
// Update the table.
|
|
UpdateTableWithReadBarrier();
|
|
|
|
// And try again.
|
|
return RemoveLocked(self, obj, tag);
|
|
}
|
|
|
|
// Not in here.
|
|
return false;
|
|
}
|
|
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::Set(art::ObjPtr<art::mirror::Object> obj, T new_tag) {
|
|
art::Thread* self = art::Thread::Current();
|
|
art::MutexLock mu(self, allow_disallow_lock_);
|
|
Wait(self);
|
|
|
|
return SetLocked(self, obj, new_tag);
|
|
}
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::SetLocked(art::ObjPtr<art::mirror::Object> obj, T new_tag) {
|
|
art::Thread* self = art::Thread::Current();
|
|
allow_disallow_lock_.AssertHeld(self);
|
|
Wait(self);
|
|
|
|
return SetLocked(self, obj, new_tag);
|
|
}
|
|
|
|
template <typename T>
|
|
bool JvmtiWeakTable<T>::SetLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T new_tag) {
|
|
auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
|
|
if (it != tagged_objects_.end()) {
|
|
it->second = new_tag;
|
|
return true;
|
|
}
|
|
|
|
if (art::kUseReadBarrier && self->GetIsGcMarking() && !update_since_last_sweep_) {
|
|
// Under concurrent GC, there is a window between moving objects and sweeping of system
|
|
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
|
|
// but still have from-space pointers in the table. Explicitly update the table once.
|
|
// Note: this will keep *all* objects in the table live, but should be a rare occurrence.
|
|
|
|
// Update the table.
|
|
UpdateTableWithReadBarrier();
|
|
|
|
// And try again.
|
|
return SetLocked(self, obj, new_tag);
|
|
}
|
|
|
|
// New element.
|
|
auto insert_it = tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(obj), new_tag);
|
|
DCHECK(insert_it.second);
|
|
return false;
|
|
}
|
|
|
|
template <typename T>
|
|
void JvmtiWeakTable<T>::Sweep(art::IsMarkedVisitor* visitor) {
|
|
if (DoesHandleNullOnSweep()) {
|
|
SweepImpl<true>(visitor);
|
|
} else {
|
|
SweepImpl<false>(visitor);
|
|
}
|
|
|
|
// Under concurrent GC, there is a window between moving objects and sweeping of system
|
|
// weaks in which mutators are active. We may receive a to-space object pointer in obj,
|
|
// but still have from-space pointers in the table. We explicitly update the table then
|
|
// to ensure we compare against to-space pointers. But we want to do this only once. Once
|
|
// sweeping is done, we know all objects are to-space pointers until the next GC cycle,
|
|
// so we re-enable the explicit update for the next marking.
|
|
update_since_last_sweep_ = false;
|
|
}
|
|
|
|
template <typename T>
|
|
template <bool kHandleNull>
|
|
void JvmtiWeakTable<T>::SweepImpl(art::IsMarkedVisitor* visitor) {
|
|
art::Thread* self = art::Thread::Current();
|
|
art::MutexLock mu(self, allow_disallow_lock_);
|
|
|
|
auto IsMarkedUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root ATTRIBUTE_UNUSED,
|
|
art::mirror::Object* original_obj) {
|
|
return visitor->IsMarked(original_obj);
|
|
};
|
|
|
|
UpdateTableWith<decltype(IsMarkedUpdater),
|
|
kHandleNull ? kCallHandleNull : kRemoveNull>(IsMarkedUpdater);
|
|
}
|
|
|
|
template <typename T>
|
|
template <typename Updater, typename JvmtiWeakTable<T>::TableUpdateNullTarget kTargetNull>
|
|
ALWAYS_INLINE inline void JvmtiWeakTable<T>::UpdateTableWith(Updater& updater) {
|
|
// We optimistically hope that elements will still be well-distributed when re-inserting them.
|
|
// So play with the map mechanics, and postpone rehashing. This avoids the need of a side
|
|
// vector and two passes.
|
|
float original_max_load_factor = tagged_objects_.max_load_factor();
|
|
tagged_objects_.max_load_factor(std::numeric_limits<float>::max());
|
|
// For checking that a max load-factor actually does what we expect.
|
|
size_t original_bucket_count = tagged_objects_.bucket_count();
|
|
|
|
for (auto it = tagged_objects_.begin(); it != tagged_objects_.end();) {
|
|
DCHECK(!it->first.IsNull());
|
|
art::mirror::Object* original_obj = it->first.template Read<art::kWithoutReadBarrier>();
|
|
art::mirror::Object* target_obj = updater(it->first, original_obj);
|
|
if (original_obj != target_obj) {
|
|
if (kTargetNull == kIgnoreNull && target_obj == nullptr) {
|
|
// Ignore null target, don't do anything.
|
|
} else {
|
|
T tag = it->second;
|
|
it = tagged_objects_.erase(it);
|
|
if (target_obj != nullptr) {
|
|
tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(target_obj), tag);
|
|
DCHECK_EQ(original_bucket_count, tagged_objects_.bucket_count());
|
|
} else if (kTargetNull == kCallHandleNull) {
|
|
HandleNullSweep(tag);
|
|
}
|
|
continue; // Iterator was implicitly updated by erase.
|
|
}
|
|
}
|
|
it++;
|
|
}
|
|
|
|
tagged_objects_.max_load_factor(original_max_load_factor);
|
|
// TODO: consider rehash here.
|
|
}
|
|
|
|
template <typename T>
|
|
template <typename Storage, class Allocator>
|
|
struct JvmtiWeakTable<T>::ReleasableContainer {
|
|
using allocator_type = Allocator;
|
|
|
|
explicit ReleasableContainer(const allocator_type& alloc, size_t reserve = 10)
|
|
: allocator(alloc),
|
|
data(reserve > 0 ? allocator.allocate(reserve) : nullptr),
|
|
size(0),
|
|
capacity(reserve) {
|
|
}
|
|
|
|
~ReleasableContainer() {
|
|
if (data != nullptr) {
|
|
allocator.deallocate(data, capacity);
|
|
capacity = 0;
|
|
size = 0;
|
|
}
|
|
}
|
|
|
|
Storage* Release() {
|
|
Storage* tmp = data;
|
|
|
|
data = nullptr;
|
|
size = 0;
|
|
capacity = 0;
|
|
|
|
return tmp;
|
|
}
|
|
|
|
void Resize(size_t new_capacity) {
|
|
CHECK_GT(new_capacity, capacity);
|
|
|
|
Storage* tmp = allocator.allocate(new_capacity);
|
|
DCHECK(tmp != nullptr);
|
|
if (data != nullptr) {
|
|
memcpy(tmp, data, sizeof(Storage) * size);
|
|
}
|
|
Storage* old = data;
|
|
data = tmp;
|
|
allocator.deallocate(old, capacity);
|
|
capacity = new_capacity;
|
|
}
|
|
|
|
void Pushback(const Storage& elem) {
|
|
if (size == capacity) {
|
|
size_t new_capacity = 2 * capacity + 1;
|
|
Resize(new_capacity);
|
|
}
|
|
data[size++] = elem;
|
|
}
|
|
|
|
Allocator allocator;
|
|
Storage* data;
|
|
size_t size;
|
|
size_t capacity;
|
|
};
|
|
|
|
template <typename T>
|
|
jvmtiError JvmtiWeakTable<T>::GetTaggedObjects(jvmtiEnv* jvmti_env,
|
|
jint tag_count,
|
|
const T* tags,
|
|
jint* count_ptr,
|
|
jobject** object_result_ptr,
|
|
T** tag_result_ptr) {
|
|
if (tag_count < 0) {
|
|
return ERR(ILLEGAL_ARGUMENT);
|
|
}
|
|
if (tag_count > 0) {
|
|
for (size_t i = 0; i != static_cast<size_t>(tag_count); ++i) {
|
|
if (tags[i] == 0) {
|
|
return ERR(ILLEGAL_ARGUMENT);
|
|
}
|
|
}
|
|
}
|
|
if (tags == nullptr) {
|
|
return ERR(NULL_POINTER);
|
|
}
|
|
if (count_ptr == nullptr) {
|
|
return ERR(NULL_POINTER);
|
|
}
|
|
|
|
art::Thread* self = art::Thread::Current();
|
|
art::MutexLock mu(self, allow_disallow_lock_);
|
|
Wait(self);
|
|
|
|
art::JNIEnvExt* jni_env = self->GetJniEnv();
|
|
|
|
constexpr size_t kDefaultSize = 10;
|
|
size_t initial_object_size;
|
|
size_t initial_tag_size;
|
|
if (tag_count == 0) {
|
|
initial_object_size = (object_result_ptr != nullptr) ? tagged_objects_.size() : 0;
|
|
initial_tag_size = (tag_result_ptr != nullptr) ? tagged_objects_.size() : 0;
|
|
} else {
|
|
initial_object_size = initial_tag_size = kDefaultSize;
|
|
}
|
|
JvmtiAllocator<void> allocator(jvmti_env);
|
|
ReleasableContainer<jobject, JvmtiAllocator<jobject>> selected_objects(allocator,
|
|
initial_object_size);
|
|
ReleasableContainer<T, JvmtiAllocator<T>> selected_tags(allocator, initial_tag_size);
|
|
|
|
size_t count = 0;
|
|
for (auto& pair : tagged_objects_) {
|
|
bool select;
|
|
if (tag_count > 0) {
|
|
select = false;
|
|
for (size_t i = 0; i != static_cast<size_t>(tag_count); ++i) {
|
|
if (tags[i] == pair.second) {
|
|
select = true;
|
|
break;
|
|
}
|
|
}
|
|
} else {
|
|
select = true;
|
|
}
|
|
|
|
if (select) {
|
|
art::ObjPtr<art::mirror::Object> obj = pair.first.template Read<art::kWithReadBarrier>();
|
|
if (obj != nullptr) {
|
|
count++;
|
|
if (object_result_ptr != nullptr) {
|
|
selected_objects.Pushback(jni_env->AddLocalReference<jobject>(obj));
|
|
}
|
|
if (tag_result_ptr != nullptr) {
|
|
selected_tags.Pushback(pair.second);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (object_result_ptr != nullptr) {
|
|
*object_result_ptr = selected_objects.Release();
|
|
}
|
|
if (tag_result_ptr != nullptr) {
|
|
*tag_result_ptr = selected_tags.Release();
|
|
}
|
|
*count_ptr = static_cast<jint>(count);
|
|
return ERR(NONE);
|
|
}
|
|
|
|
template <typename T>
|
|
art::ObjPtr<art::mirror::Object> JvmtiWeakTable<T>::Find(T tag) {
|
|
art::Thread* self = art::Thread::Current();
|
|
art::MutexLock mu(self, allow_disallow_lock_);
|
|
Wait(self);
|
|
|
|
for (auto& pair : tagged_objects_) {
|
|
if (tag == pair.second) {
|
|
art::ObjPtr<art::mirror::Object> obj = pair.first.template Read<art::kWithReadBarrier>();
|
|
if (obj != nullptr) {
|
|
return obj;
|
|
}
|
|
}
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
} // namespace openjdkjvmti
|
|
|
|
#endif // ART_OPENJDKJVMTI_JVMTI_WEAK_TABLE_INL_H_
|