/art/runtime/ |
D | non_debuggable_classes.cc | 31 Thread* self = Thread::Current(); in AddNonDebuggableClass() local 32 JNIEnvExt* env = self->GetJniEnv(); in AddNonDebuggableClass() 33 ObjPtr<mirror::Class> mirror_klass(self->DecodeJObject(klass)->AsClass()); in AddNonDebuggableClass() 35 if (self->DecodeJObject(c)->AsClass() == mirror_klass) { in AddNonDebuggableClass()
|
D | monitor.h | 88 static void Notify(Thread* self, ObjPtr<mirror::Object> obj) in Notify() argument 90 DoNotify(self, obj, false); in Notify() 92 static void NotifyAll(Thread* self, ObjPtr<mirror::Object> obj) in NotifyAll() argument 94 DoNotify(self, obj, true); in NotifyAll() 99 static void Wait(Thread* self, 152 static void InflateThinLocked(Thread* self, Handle<mirror::Object> obj, LockWord lock_word, 158 static bool Deflate(Thread* self, ObjPtr<mirror::Object> obj) 176 Monitor(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code) 178 Monitor(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code, MonitorId id) 183 bool Install(Thread* self) [all …]
|
D | linear_alloc.h | 32 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_); 33 void* AllocAlign16(Thread* self, size_t size) REQUIRES(!lock_); 36 void* Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) REQUIRES(!lock_); 40 T* AllocArray(Thread* self, size_t elements) REQUIRES(!lock_) { in AllocArray() argument 41 return reinterpret_cast<T*>(Alloc(self, elements * sizeof(T))); in AllocArray()
|
D | imtable_test.cc | 46 Thread* self = Thread::Current(); in LoadClasses() local 48 StackHandleScope<3> hs(self); in LoadClasses() 53 ObjPtr<mirror::ClassLoader>::DownCast(self->DecodeJObject(jclass_loader_a))); in LoadClasses() 55 hs.NewHandle(class_linker->FindClass(self, class_name.c_str(), h_class_loader))); in LoadClasses() 57 LOG(ERROR) << self->GetException()->Dump(); in LoadClasses() 63 ObjPtr<mirror::ClassLoader>::DownCast(self->DecodeJObject(jclass_loader_b))); in LoadClasses() 65 hs.NewHandle(class_linker->FindClass(self, class_name.c_str(), h_class_loader))); in LoadClasses() 67 LOG(ERROR) << self->GetException()->Dump(); in LoadClasses()
|
D | intern_table.cc | 92 ObjPtr<mirror::String> InternTable::LookupWeak(Thread* self, ObjPtr<mirror::String> s) { in LookupWeak() argument 93 MutexLock mu(self, *Locks::intern_table_lock_); in LookupWeak() 97 ObjPtr<mirror::String> InternTable::LookupStrong(Thread* self, ObjPtr<mirror::String> s) { in LookupStrong() argument 98 MutexLock mu(self, *Locks::intern_table_lock_); in LookupStrong() 102 ObjPtr<mirror::String> InternTable::LookupStrong(Thread* self, in LookupStrong() argument 109 MutexLock mu(self, *Locks::intern_table_lock_); in LookupStrong() 182 Thread* self = Thread::Current(); in BroadcastForNewInterns() local 183 MutexLock mu(self, *Locks::intern_table_lock_); in BroadcastForNewInterns() 184 weak_intern_condition_.Broadcast(self); in BroadcastForNewInterns() 187 void InternTable::WaitUntilAccessible(Thread* self) { in WaitUntilAccessible() argument [all …]
|
D | var_handles.cc | 30 bool VarHandleInvokeAccessorWithConversions(Thread* self, in VarHandleInvokeAccessorWithConversions() argument 38 StackHandleScope<1> hs(self); in VarHandleInvokeAccessorWithConversions() 40 var_handle->GetMethodTypeForAccessMode(self, access_mode))); in VarHandleInvokeAccessorWithConversions() 48 if (!PerformConversions(self, callsite_type, accessor_type, &getter, &setter, num_params)) { in VarHandleInvokeAccessorWithConversions() 61 bool VarHandleInvokeAccessor(Thread* self, in VarHandleInvokeAccessor() argument 83 return VarHandleInvokeAccessorWithConversions(self, in VarHandleInvokeAccessor()
|
/art/runtime/mirror/ |
D | method_type.cc | 31 ObjPtr<ObjectArray<Class>> AllocatePTypesArray(Thread* self, int count) in AllocatePTypesArray() argument 34 return ObjectArray<Class>::Alloc(self, class_array_type, count); in AllocatePTypesArray() 39 ObjPtr<MethodType> MethodType::Create(Thread* const self, in Create() argument 42 StackHandleScope<1> hs(self); in Create() 44 hs.NewHandle(ObjPtr<MethodType>::DownCast(GetClassRoot<MethodType>()->AllocObject(self)))); in Create() 58 ObjPtr<MethodType> MethodType::CloneWithoutLeadingParameter(Thread* const self, in CloneWithoutLeadingParameter() argument 60 StackHandleScope<3> hs(self); in CloneWithoutLeadingParameter() 64 Handle<ObjectArray<Class>> dst_ptypes = hs.NewHandle(AllocatePTypesArray(self, dst_ptypes_count)); in CloneWithoutLeadingParameter() 71 return Create(self, dst_rtype, dst_ptypes); in CloneWithoutLeadingParameter() 74 ObjPtr<MethodType> MethodType::CollectTrailingArguments(Thread* self, in CollectTrailingArguments() argument [all …]
|
D | object_array-alloc-inl.h | 36 inline ObjPtr<ObjectArray<T>> ObjectArray<T>::Alloc(Thread* self, in Alloc() argument 40 ObjPtr<Array> array = Array::Alloc(self, in Alloc() 54 inline ObjPtr<ObjectArray<T>> ObjectArray<T>::Alloc(Thread* self, in Alloc() argument 57 return Alloc(self, in Alloc() 65 Thread* self, in CopyOf() argument 72 ObjPtr<ObjectArray<T>> new_array = Alloc(self, h_this->GetClass(), new_length, allocator_type); in CopyOf()
|
D | field.cc | 57 ObjPtr<mirror::Field> Field::CreateFromArtField(Thread* self, in CreateFromArtField() argument 60 StackHandleScope<2> hs(self); in CreateFromArtField() 65 DCHECK(self->IsExceptionPending()); in CreateFromArtField() 70 mirror::Throwable* exception = self->GetException(); in CreateFromArtField() 74 self->ClearException(); in CreateFromArtField() 77 auto ret = hs.NewHandle(ObjPtr<Field>::DownCast(GetClassRoot<Field>()->AllocObject(self))); in CreateFromArtField() 79 self->AssertPendingOOMException(); in CreateFromArtField()
|
D | class_ext.cc | 64 bool ClassExt::ExtendObsoleteArrays(Handle<ClassExt> h_this, Thread* self, uint32_t increase) { in ExtendObsoleteArrays() argument 66 StackHandleScope<4> hs(self); in ExtendObsoleteArrays() 79 cl->AllocPointerArray(self, new_len))); in ExtendObsoleteArrays() 82 self->AssertPendingOOMException(); in ExtendObsoleteArrays() 86 ObjectArray<DexCache>::Alloc(self, in ExtendObsoleteArrays() 87 cl->FindClass(self, in ExtendObsoleteArrays() 93 self->AssertPendingOOMException(); in ExtendObsoleteArrays() 117 ObjPtr<ClassExt> ClassExt::Alloc(Thread* self) { in Alloc() argument 118 return ObjPtr<ClassExt>::DownCast(GetClassRoot<ClassExt>()->AllocObject(self)); in Alloc()
|
/art/runtime/gc/space/ |
D | large_object_space_test.cc | 38 Thread* const self = Thread::Current(); in LargeObjectTest() local 63 mirror::Object* obj = los->Alloc(self, request_size, &allocation_size, nullptr, in LargeObjectTest() 86 ASSERT_FALSE(los->IsZygoteLargeObject(self, obj)); in LargeObjectTest() 88 los->SetAllLargeObjectsAsZygoteObjects(self, /*set_mark_bit=*/ false); in LargeObjectTest() 91 ASSERT_TRUE(los->IsZygoteLargeObject(self, obj)); in LargeObjectTest() 115 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr, in LargeObjectTest() 131 void Run(Thread* self) override { in Run() argument 134 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr, in Run() 139 los_->Free(self, ptr); in Run() 163 Thread* self = Thread::Current(); in RaceTest() local [all …]
|
D | bump_pointer_space.cc | 112 Thread* self = Thread::Current(); in RevokeAllThreadLocalBuffers() local 113 MutexLock mu(self, *Locks::runtime_shutdown_lock_); in RevokeAllThreadLocalBuffers() 114 MutexLock mu2(self, *Locks::thread_list_lock_); in RevokeAllThreadLocalBuffers() 132 Thread* self = Thread::Current(); in AssertAllThreadLocalBuffersAreRevoked() local 133 MutexLock mu(self, *Locks::runtime_shutdown_lock_); in AssertAllThreadLocalBuffersAreRevoked() 134 MutexLock mu2(self, *Locks::thread_list_lock_); in AssertAllThreadLocalBuffersAreRevoked() 173 Thread* self = Thread::Current(); in GetBytesAllocated() local 174 MutexLock mu(self, *Locks::runtime_shutdown_lock_); in GetBytesAllocated() 175 MutexLock mu2(self, *Locks::thread_list_lock_); in GetBytesAllocated() 191 Thread* self = Thread::Current(); in GetObjectsAllocated() local [all …]
|
D | rosalloc_space.h | 56 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 59 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, in Alloc() argument 61 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, in Alloc() 64 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, in AllocThreadUnsafe() argument 67 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, in AllocThreadUnsafe() 73 size_t Free(Thread* self, mirror::Object* ptr) override 75 size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) override 78 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, in AllocNonvirtual() argument 81 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, in AllocNonvirtual() 84 mirror::Object* AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, in AllocNonvirtualThreadUnsafe() argument [all …]
|
/art/runtime/arch/arm/ |
D | quick_entrypoints_cc_arm.cc | 24 Thread* self, JValue* result, uint32_t, uint32_t*, 29 Thread* self, JValue* result, const char* shorty) { in quick_invoke_reg_setup() argument 93 art_quick_invoke_stub_internal(method, args, args_size, self, result, result_in_float, in quick_invoke_reg_setup() 100 Thread* self, JValue* result, const char* shorty) { in art_quick_invoke_stub() argument 101 quick_invoke_reg_setup<false>(method, args, args_size, self, result, shorty); in art_quick_invoke_stub() 107 uint32_t args_size, Thread* self, JValue* result, in art_quick_invoke_static_stub() argument 109 quick_invoke_reg_setup<true>(method, args, args_size, self, result, shorty); in art_quick_invoke_static_stub()
|
/art/runtime/entrypoints/quick/ |
D | quick_trampoline_entrypoints.cc | 589 Thread* self = Thread::Current(); in HandleDeoptimization() local 604 QuickExceptionHandler::DumpFramesWithType(self, true); in HandleDeoptimization() 610 self->PopDeoptimizationContext(/* out */ result, in HandleDeoptimization() 616 self->PushManagedStackFragment(fragment); in HandleDeoptimization() 631 EntireStackVisitor esv(self); in HandleDeoptimization() 638 self->SetException(pending_exception); in HandleDeoptimization() 640 interpreter::EnterInterpreterFromDeoptimize(self, in HandleDeoptimization() 647 extern "C" uint64_t artQuickToInterpreterBridge(ArtMethod* method, Thread* self, ArtMethod** sp) in artQuickToInterpreterBridge() argument 651 ScopedQuickEntrypointChecks sqec(self); in artQuickToInterpreterBridge() 659 ShadowFrame* deopt_frame = self->PopStackedShadowFrame( in artQuickToInterpreterBridge() [all …]
|
/art/runtime/gc/ |
D | task_processor.h | 57 void AddTask(Thread* self, HeapTask* task) REQUIRES(!lock_); 58 HeapTask* GetTask(Thread* self) REQUIRES(!lock_); 59 void Start(Thread* self) REQUIRES(!lock_); 62 void Stop(Thread* self) REQUIRES(!lock_); 63 void RunAllTasks(Thread* self) REQUIRES(!lock_); 65 void UpdateTargetRunTime(Thread* self, HeapTask* target_time, uint64_t new_target_time)
|
D | scoped_gc_critical_section.cc | 50 ScopedGCCriticalSection::ScopedGCCriticalSection(Thread* self, in ScopedGCCriticalSection() argument 53 : critical_section_(self, "ScopedGCCriticalSection") { in ScopedGCCriticalSection() 62 Thread* self, in ScopedInterruptibleGCCriticalSection() argument 64 CollectorType type) : self_(self) { in ScopedInterruptibleGCCriticalSection() 65 DCHECK(self != nullptr); in ScopedInterruptibleGCCriticalSection()
|
D | reference_processor.h | 60 void BroadcastForSlowPath(Thread* self); 62 ObjPtr<mirror::Object> GetReferent(Thread* self, ObjPtr<mirror::Reference> reference) 66 SelfDeletingTask* CollectClearedReferences(Thread* self) REQUIRES(!Locks::mutator_lock_); 85 void DisableSlowPath(Thread* self) REQUIRES(Locks::reference_processor_lock_) 90 void StartPreservingReferences(Thread* self) REQUIRES(!Locks::reference_processor_lock_); 91 void StopPreservingReferences(Thread* self) REQUIRES(!Locks::reference_processor_lock_); 93 void WaitUntilDoneProcessingReferences(Thread* self)
|
D | heap.cc | 875 void Heap::IncrementDisableMovingGC(Thread* self) { in IncrementDisableMovingGC() argument 878 ScopedThreadStateChange tsc(self, kWaitingForGcToComplete); in IncrementDisableMovingGC() 879 MutexLock mu(self, *gc_complete_lock_); in IncrementDisableMovingGC() 882 WaitForGcToCompleteLocked(kGcCauseDisableMovingGc, self); in IncrementDisableMovingGC() 886 void Heap::DecrementDisableMovingGC(Thread* self) { in DecrementDisableMovingGC() argument 887 MutexLock mu(self, *gc_complete_lock_); in DecrementDisableMovingGC() 892 void Heap::IncrementDisableThreadFlip(Thread* self) { in IncrementDisableThreadFlip() argument 895 bool is_nested = self->GetDisableThreadFlipCount() > 0; in IncrementDisableThreadFlip() 896 self->IncrementDisableThreadFlipCount(); in IncrementDisableThreadFlip() 902 ScopedThreadStateChange tsc(self, kWaitingForGcThreadFlip); in IncrementDisableThreadFlip() [all …]
|
/art/openjdkjvmti/ |
D | jvmti_weak_table.h | 84 art::Thread* self = art::Thread::Current(); in GetTag() local 85 art::MutexLock mu(self, allow_disallow_lock_); in GetTag() 86 Wait(self); in GetTag() 88 return GetTagLocked(self, obj, result); in GetTag() 93 art::Thread* self = art::Thread::Current(); in GetTagLocked() local 94 allow_disallow_lock_.AssertHeld(self); in GetTagLocked() 95 Wait(self); in GetTagLocked() 97 return GetTagLocked(self, obj, result); in GetTagLocked() 135 bool SetLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T tag) 140 bool RemoveLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, /* out */ T* tag) [all …]
|
D | ti_stack.cc | 168 void Run(art::Thread* self) override REQUIRES_SHARED(art::Locks::mutator_lock_) { in Run() 172 auto visitor = MakeStackTraceVisitor(self, start_input, stop_input, frames_fn); in Run() 235 void Run(art::Thread* self) override REQUIRES_SHARED(art::Locks::mutator_lock_) { in Run() 240 auto visitor = MakeStackTraceVisitor(self, start_input, stop_input, frames_fn); in Run() 332 art::Thread* self = art::Thread::Current(); in Run() local 333 Work(thread, self); in Run() 334 barrier.Pass(self); in Run() 337 void Work(art::Thread* thread, art::Thread* self) in Work() 345 std::vector<jvmtiFrameInfo>* thread_frames = data->GetFrameStorageFor(self, thread); in Work() 372 art::Thread* self = art::Thread::Current(); in RunCheckpointAndWait() local [all …]
|
/art/test/970-iface-super-resolution-gen/util-src/ |
D | generate_java.py | 40 def __init__(self, inner): argument 41 self.inner = inner 43 def get_name(self): argument 44 return self.inner.get_name() 46 def __str__(self): argument 48 for line in str(self.inner).splitlines(keepends = True):
|
/art/runtime/gc/accounting/ |
D | mod_union_table_test.cc | 51 Thread* self, space::ContinuousMemMapAllocSpace* space, size_t component_count) in AllocObjectArray() argument 53 auto* klass = GetObjectArrayClass(self, space); in AllocObjectArray() 57 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated)); in AllocObjectArray() 72 mirror::Class* GetObjectArrayClass(Thread* self, space::ContinuousMemMapAllocSpace* space) in GetObjectArrayClass() argument 84 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated, in GetObjectArrayClass() 179 Thread* const self = Thread::Current(); in RunTest() local 180 ScopedObjectAccess soa(self); in RunTest() 191 ScopedThreadSuspension sts(self, kSuspended); in RunTest() 199 auto* obj1 = AllocObjectArray(self, space, CardTable::kCardSize); in RunTest() 201 auto* obj2 = AllocObjectArray(self, space, CardTable::kCardSize); in RunTest() [all …]
|
/art/runtime/gc/collector/ |
D | concurrent_copying.cc | 124 Thread* self = Thread::Current(); in ConcurrentCopying() local 126 ReaderMutexLock mu(self, *Locks::heap_bitmap_lock_); in ConcurrentCopying() 134 MutexLock mu(self, mark_stack_lock_); in ConcurrentCopying() 158 Thread* const self = Thread::Current(); in MarkHeapReference() local 165 mirror::Object* to_ref = Mark(self, from_ref); in MarkHeapReference() 178 field->Assign(Mark(self, field->AsMirrorPtr())); in MarkHeapReference() 190 Thread* self = Thread::Current(); in RunPhases() local 191 thread_running_gc_ = self; in RunPhases() 192 Locks::mutator_lock_->AssertNotHeld(self); in RunPhases() 194 ReaderMutexLock mu(self, *Locks::mutator_lock_); in RunPhases() [all …]
|
/art/runtime/jit/ |
D | jit.cc | 294 Thread* self, in CompileMethod() argument 337 method_to_compile, self, compilation_kind, prejit, region)) { in CompileMethod() 344 bool success = jit_compiler_->CompileMethod(self, region, method_to_compile, compilation_kind); in CompileMethod() 345 code_cache_->DoneCompiling(method_to_compile, self, compilation_kind); in CompileMethod() 352 if (self->IsExceptionPending()) { in CompileMethod() 353 mirror::Throwable* exception = self->GetException(); in CompileMethod() 370 Thread* self = Thread::Current(); in DeleteThreadPool() local 382 pool->StopWorkers(self); in DeleteThreadPool() 383 pool->RemoveAllTasks(self); in DeleteThreadPool() 388 pool->Wait(self, false, false); in DeleteThreadPool() [all …]
|