/art/runtime/ |
D | monitor_pool.h | 45 static Monitor* CreateMonitor(Thread* self, in CreateMonitor() argument 51 Monitor* mon = new Monitor(self, owner, obj, hash_code); in CreateMonitor() 55 return GetMonitorPool()->CreateMonitorInPool(self, owner, obj, hash_code); in CreateMonitor() 59 static void ReleaseMonitor(Thread* self, Monitor* monitor) { in ReleaseMonitor() argument 61 UNUSED(self); in ReleaseMonitor() 64 GetMonitorPool()->ReleaseMonitorToPool(self, monitor); in ReleaseMonitor() 68 static void ReleaseMonitors(Thread* self, MonitorList::Monitors* monitors) { in ReleaseMonitors() argument 70 UNUSED(self); in ReleaseMonitors() 73 GetMonitorPool()->ReleaseMonitorsToPool(self, monitors); in ReleaseMonitors() 93 static MonitorId ComputeMonitorId(Monitor* mon, Thread* self) { in ComputeMonitorId() argument [all …]
|
D | barrier.h | 53 void Pass(Thread* self) REQUIRES(!GetLock()); 56 void Wait(Thread* self) REQUIRES(!GetLock()); 67 void Increment(Thread* self, int delta) REQUIRES(!GetLock()); 71 bool Increment(Thread* self, int delta, uint32_t timeout_ms) REQUIRES(!GetLock()); 75 void Init(Thread* self, int count) REQUIRES(!GetLock()); 77 int GetCount(Thread* self) REQUIRES(!GetLock()); 80 void SetCountLocked(Thread* self, int count) REQUIRES(GetLock());
|
D | method_handles.cc | 336 Thread* self, in ConvertAndCopyArgumentsFromCallerFrame() argument 355 return PerformConversions<ShadowFrameGetter, ShadowFrameSetter>(self, in ConvertAndCopyArgumentsFromCallerFrame() 423 Thread* self, in MethodHandleInvokeMethod() argument 463 self, new_shadow_frame, StackedShadowFrameType::kShadowFrameUnderConstruction); in MethodHandleInvokeMethod() 485 if (!emulated_stack_frame->WriteToShadowFrame(self, in MethodHandleInvokeMethod() 489 DCHECK(self->IsExceptionPending()); in MethodHandleInvokeMethod() 498 if (!ConvertAndCopyArgumentsFromCallerFrame(self, in MethodHandleInvokeMethod() 505 DCHECK(self->IsExceptionPending()); in MethodHandleInvokeMethod() 515 PerformCall(self, in MethodHandleInvokeMethod() 522 if (self->IsExceptionPending()) { in MethodHandleInvokeMethod() [all …]
|
D | thread_pool.h | 35 virtual void Run(Thread* self) = 0; 41 void Run(Thread* self) override { in Run() argument 42 func_(self); in Run() 67 void Run(Thread* self) override { in Run() argument 68 func_(self); in Run() 118 void StartWorkers(Thread* self) REQUIRES(!task_queue_lock_); 121 void StopWorkers(Thread* self) REQUIRES(!task_queue_lock_); 125 void AddTask(Thread* self, Task* task) REQUIRES(!task_queue_lock_); 128 void RemoveAllTasks(Thread* self) REQUIRES(!task_queue_lock_); 150 void Wait(Thread* self, bool do_work, bool may_hold_locks) REQUIRES(!task_queue_lock_); [all …]
|
D | read_barrier-inl.h | 42 Thread* const self = Thread::Current(); in Barrier() local 43 if (self != nullptr) { in Barrier() 44 CHECK_EQ(self->GetDebugDisallowReadBarrierCount(), 0u); in Barrier() 113 Thread* const self = Thread::Current(); in BarrierForRoot() local 114 if (self != nullptr) { in BarrierForRoot() 115 CHECK_EQ(self->GetDebugDisallowReadBarrierCount(), 0u); in BarrierForRoot() 120 Thread* self = Thread::Current(); in BarrierForRoot() local 121 if (self != nullptr && self->GetIsGcMarking()) { in BarrierForRoot() 130 Thread* self = Thread::Current(); in BarrierForRoot() local 131 if (self != nullptr && in BarrierForRoot() [all …]
|
D | signal_catcher.cc | 81 Thread* self = Thread::Current(); in SignalCatcher() local 82 MutexLock mu(self, lock_); in SignalCatcher() 84 cond_.Wait(self); in SignalCatcher() 151 int SignalCatcher::WaitForSignal(Thread* self, SignalSet& signals) { in WaitForSignal() argument 152 ScopedThreadStateChange tsc(self, kWaitingInMainSignalCatcherLoop); in WaitForSignal() 162 LOG(INFO) << *self << ": reacting to signal " << signal_number; in WaitForSignal() 179 Thread* self = Thread::Current(); in Run() local 180 DCHECK_NE(self->GetState(), kRunnable); in Run() 182 MutexLock mu(self, signal_catcher->lock_); in Run() 183 signal_catcher->thread_ = self; in Run() [all …]
|
D | class_linker.cc | 165 Thread* self = Thread::Current(); in ThrowNoClassDefFoundError() local 166 self->ThrowNewExceptionV("Ljava/lang/NoClassDefFoundError;", fmt, args); in ThrowNoClassDefFoundError() 170 static bool HasInitWithString(Thread* self, ClassLinker* class_linker, const char* descriptor) in HasInitWithString() argument 172 ArtMethod* method = self->GetCurrentMethod(nullptr); in HasInitWithString() 173 StackHandleScope<1> hs(self); in HasInitWithString() 176 ObjPtr<mirror::Class> exception_class = class_linker->FindClass(self, descriptor, class_loader); in HasInitWithString() 180 CHECK(self->IsExceptionPending()); in HasInitWithString() 181 self->ClearException(); in HasInitWithString() 201 static void HandleEarlierVerifyError(Thread* self, in HandleEarlierVerifyError() argument 207 self->AssertNoPendingException(); in HandleEarlierVerifyError() [all …]
|
D | linear_alloc.cc | 26 void* LinearAlloc::Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) { in Realloc() argument 27 MutexLock mu(self, lock_); in Realloc() 31 void* LinearAlloc::Alloc(Thread* self, size_t size) { in Alloc() argument 32 MutexLock mu(self, lock_); in Alloc() 36 void* LinearAlloc::AllocAlign16(Thread* self, size_t size) { in AllocAlign16() argument 37 MutexLock mu(self, lock_); in AllocAlign16()
|
/art/runtime/jit/ |
D | jit-inl.h | 30 inline bool Jit::ShouldUsePriorityThreadWeight(Thread* self) { in ShouldUsePriorityThreadWeight() argument 31 return self->IsJitSensitiveThread() && Runtime::Current()->InJankPerceptibleProcessState(); in ShouldUsePriorityThreadWeight() 34 inline void Jit::AddSamples(Thread* self, in AddSamples() argument 38 if (Jit::ShouldUsePriorityThreadWeight(self)) { in AddSamples() 50 if (!MaybeCompileMethod(self, method, old_count, new_count, with_backedges)) { in AddSamples() 55 if (!MaybeCompileMethod(self, method, old_batch, new_batch, with_backedges)) { in AddSamples()
|
/art/runtime/entrypoints/quick/ |
D | quick_dexcache_entrypoints.cc | 135 extern "C" mirror::Class* artInitializeStaticStorageFromCode(mirror::Class* klass, Thread* self) in artInitializeStaticStorageFromCode() argument 140 ScopedQuickEntrypointChecks sqec(self); in artInitializeStaticStorageFromCode() 143 StackHandleScope<1> hs(self); in artInitializeStaticStorageFromCode() 146 self, h_klass, /* can_init_fields= */ true, /* can_init_parents= */ true); in artInitializeStaticStorageFromCode() 153 extern "C" mirror::Class* artResolveTypeFromCode(uint32_t type_idx, Thread* self) in artResolveTypeFromCode() argument 156 ScopedQuickEntrypointChecks sqec(self); in artResolveTypeFromCode() 158 self, CalleeSaveType::kSaveEverythingForClinit); in artResolveTypeFromCode() 162 self, in artResolveTypeFromCode() 171 extern "C" mirror::Class* artResolveTypeAndVerifyAccessFromCode(uint32_t type_idx, Thread* self) in artResolveTypeAndVerifyAccessFromCode() argument 174 ScopedQuickEntrypointChecks sqec(self); in artResolveTypeAndVerifyAccessFromCode() [all …]
|
D | quick_alloc_entrypoints.cc | 40 Thread* self) REQUIRES_SHARED(Locks::mutator_lock_) { in artAllocObjectFromCode() argument 41 ScopedQuickEntrypointChecks sqec(self); in artAllocObjectFromCode() 48 if (LIKELY(byte_count < self->TlabSize())) { in artAllocObjectFromCode() 51 mirror::Object* obj = self->AllocTlab(byte_count); in artAllocObjectFromCode() 62 return AllocObjectFromCodeInitialized<kInstrumented>(klass, self, allocator_type).Ptr(); in artAllocObjectFromCode() 64 return AllocObjectFromCodeResolved<kInstrumented>(klass, self, allocator_type).Ptr(); in artAllocObjectFromCode() 66 return AllocObjectFromCode<kInstrumented>(klass, self, allocator_type).Ptr(); in artAllocObjectFromCode() 72 mirror::Class* klass, Thread* self) \ 74 return artAllocObjectFromCode<false, true, instrumented_bool, allocator_type>(klass, self); \ 77 mirror::Class* klass, Thread* self) \ [all …]
|
/art/runtime/interpreter/mterp/ |
D | nterp.cc | 103 inline void UpdateCache(Thread* self, uint16_t* dex_pc_ptr, T value) { in UpdateCache() argument 108 if (self->GetWeakRefAccessEnabled()) { in UpdateCache() 109 self->GetInterpreterCache()->Set(dex_pc_ptr, value); in UpdateCache() 114 inline void UpdateCache(Thread* self, uint16_t* dex_pc_ptr, T* value) { in UpdateCache() argument 115 UpdateCache(self, dex_pc_ptr, reinterpret_cast<size_t>(value)); in UpdateCache() 158 extern "C" size_t NterpGetMethod(Thread* self, ArtMethod* caller, uint16_t* dex_pc_ptr) in NterpGetMethod() argument 231 self, method_index, caller, invoke_type) in NterpGetMethod() 233 self, method_index, caller, invoke_type); in NterpGetMethod() 235 DCHECK(self->IsExceptionPending()); in NterpGetMethod() 286 UpdateCache(self, dex_pc_ptr, resolved_method->GetImtIndex()); in NterpGetMethod() [all …]
|
/art/runtime/gc/collector/ |
D | concurrent_copying-inl.h | 36 Thread* const self, in MarkUnevacFromSpaceRegion() argument 53 PushOntoMarkStack(self, ref); in MarkUnevacFromSpaceRegion() 89 PushOntoMarkStack(self, ref); in MarkUnevacFromSpaceRegion() 95 inline mirror::Object* ConcurrentCopying::MarkImmuneSpace(Thread* const self, in MarkImmuneSpace() argument 103 if (self == thread_running_gc_) { in MarkImmuneSpace() 119 MutexLock mu(self, immune_gray_stack_lock_); in MarkImmuneSpace() 127 inline mirror::Object* ConcurrentCopying::Mark(Thread* const self, in Mark() argument 139 DCHECK_EQ(self, thread_running_gc_); in Mark() 165 to_ref = Copy(self, from_ref, holder, offset); in Mark() 180 return MarkUnevacFromSpaceRegion(self, from_ref, region_space_bitmap_); in Mark() [all …]
|
/art/openjdkjvmti/ |
D | events.cc | 273 art::Thread* self, in RunEventCallback() argument 277 ScopedLocalRef<jthread> thread_jni(jnienv, AddLocalRef<jthread>(jnienv, self->GetPeer())); in RunEventCallback() 278 handler->DispatchEvent<kEvent>(self, in RunEventCallback() 300 art::Thread* self = art::Thread::Current(); in DdmPublishChunk() local 302 self, in DdmPublishChunk() 319 void ObjectAllocated(art::Thread* self, art::ObjPtr<art::mirror::Object>* obj, size_t byte_count) in ObjectAllocated() argument 321 DCHECK_EQ(self, art::Thread::Current()); in ObjectAllocated() 324 art::StackHandleScope<1> hs(self); in ObjectAllocated() 333 art::JNIEnvExt* jni_env = self->GetJniEnv(); in ObjectAllocated() 340 self, in ObjectAllocated() [all …]
|
D | ti_class_loader.cc | 63 bool ClassLoaderHelper::AddToClassLoader(art::Thread* self, in AddToClassLoader() argument 66 art::ScopedObjectAccessUnchecked soa(self); in AddToClassLoader() 67 art::StackHandleScope<3> hs(self); in AddToClassLoader() 69 art::Runtime::Current()->GetClassLinker()->AppendToBootClassPath(self, dex_file); in AddToClassLoader() 73 hs.NewHandle(FindSourceDexFileObject(self, loader))); in AddToClassLoader() 79 AllocateNewDexFileCookie(self, old_cookie, dex_file))); in AddToClassLoader() 116 art::Thread* self, in AllocateNewDexFileCookie() argument 119 art::StackHandleScope<1> hs(self); in AllocateNewDexFileCookie() 123 hs.NewHandle(art::mirror::LongArray::Alloc(self, cookie->GetLength() + 1))); in AllocateNewDexFileCookie() 125 self->AssertPendingOOMException(); in AllocateNewDexFileCookie() [all …]
|
D | alloc_manager.h | 60 void ObjectAllocated(art::Thread* self, 64 void PreObjectAllocated(art::Thread* self, 77 virtual void ObjectAllocated(art::Thread* self, 89 void PauseAllocations(art::Thread* self) REQUIRES_SHARED(art::Locks::mutator_lock_); 90 void ResumeAllocations(art::Thread* self) REQUIRES_SHARED(art::Locks::mutator_lock_); 92 void EnableAllocationCallback(art::Thread* self) REQUIRES_SHARED(art::Locks::mutator_lock_); 93 void DisableAllocationCallback(art::Thread* self) REQUIRES_SHARED(art::Locks::mutator_lock_); 97 void PauseForAllocation(art::Thread* self, T msg) REQUIRES_SHARED(art::Locks::mutator_lock_); 98 void IncrListenerInstall(art::Thread* self) REQUIRES_SHARED(art::Locks::mutator_lock_); 99 void DecrListenerInstall(art::Thread* self) REQUIRES_SHARED(art::Locks::mutator_lock_);
|
D | transform.cc | 94 art::Thread* self = art::Thread::Current(); in Action() local 96 if (UNLIKELY(uninitialized_class_definitions_lock_.IsExclusiveHeld(self))) { in Action() 114 art::MutexLock mu(self, uninitialized_class_definitions_lock_); in Action() 139 if (LIKELY(self != nullptr)) { in Action() 140 CHECK_EQ(self->GetState(), art::ThreadState::kNative) in Action() 149 art::MutexLock mu(self, uninitialized_class_definitions_lock_); in Action() 155 class_definition_initialized_cond_.Broadcast(self); in Action() 257 EventHandler* event_handler, art::Thread* self, /*in-out*/ArtClassDefinition* def); 260 EventHandler* event_handler, art::Thread* self, /*in-out*/ArtClassDefinition* def); 263 EventHandler* event_handler, art::Thread* self, /*in-out*/ArtClassDefinition* def); [all …]
|
/art/tools/checker/common/ |
D | immutables.py | 16 def __setitem__(self, key, value): argument 19 def __delitem__(self, key): argument 22 def copyWith(self, key, value): argument 23 newDict = ImmutableDict(self)
|
/art/runtime/gc/space/ |
D | rosalloc_space-inl.h | 32 inline mirror::Object* RosAllocSpace::AllocCommon(Thread* self, size_t num_bytes, in AllocCommon() argument 39 Locks::mutator_lock_->AssertExclusiveHeld(self); in AllocCommon() 42 rosalloc_->Alloc<kThreadSafe>(self, num_bytes, &rosalloc_bytes_allocated, in AllocCommon() 62 inline bool RosAllocSpace::CanAllocThreadLocal(Thread* self, size_t num_bytes) { in CanAllocThreadLocal() argument 63 return rosalloc_->CanAllocFromThreadLocalRun(self, num_bytes); in CanAllocThreadLocal() 66 inline mirror::Object* RosAllocSpace::AllocThreadLocal(Thread* self, size_t num_bytes, in AllocThreadLocal() argument 70 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated)); in AllocThreadLocal()
|
/art/test/988-method-trace/ |
D | gen_srcs.py | 169 def __init__(self, staticness, pretty_params, method, kls): argument 171 self.staticness = staticness 173 self.parameters = pretty_params 175 self.method_name = method 177 self.klass = kls 179 def __str__(self): argument 180 return "MethodInfo " + str(self.__dict__) 182 def placeholder_parameters(self): argument 199 return [ placeholder_values.get(param, object_placeholder(param)) for param in self.parameters ] 201 def placeholder_instance_value(self): argument [all …]
|
/art/runtime/interpreter/ |
D | interpreter.h | 41 extern void EnterInterpreterFromInvoke(Thread* self, ArtMethod* method, 49 extern void EnterInterpreterFromDeoptimize(Thread* self, 56 extern JValue EnterInterpreterFromEntryPoint(Thread* self, 61 void ArtInterpreterToInterpreterBridge(Thread* self, 70 void InitInterpreterTls(Thread* self); 75 bool PrevFrameWillRetry(Thread* self, const ShadowFrame& frame)
|
/art/runtime/native/ |
D | java_lang_VMClassLoader.cc | 42 Thread* self, in LookupClass() argument 48 return cl->LookupClass(self, descriptor, hash, class_loader); in LookupClass() 53 Thread* self, in FindClassInPathClassLoader() argument 59 if (cl->FindClassInBaseDexClassLoader(soa, self, descriptor, hash, class_loader, &result)) { in FindClassInPathClassLoader() 60 DCHECK(!self->IsExceptionPending()); in FindClassInPathClassLoader() 63 if (self->IsExceptionPending()) { in FindClassInPathClassLoader() 64 self->ClearException(); in FindClassInPathClassLoader() 95 Thread* self = soa.Self(); in VMClassLoader_findLoadedClass() local 97 self->DecodeJObject(WellKnownClasses::java_lang_IllegalAccessError)->AsClass(); in VMClassLoader_findLoadedClass() 99 self->DecodeJObject(WellKnownClasses::java_lang_NoClassDefFoundError)->AsClass(); in VMClassLoader_findLoadedClass() [all …]
|
/art/runtime/entrypoints/ |
D | entrypoint_utils-inl.h | 120 Thread* self, in CheckClassInitializedForObjectAlloc() argument 125 StackHandleScope<1> hs(self); in CheckClassInitializedForObjectAlloc() 136 if (!Runtime::Current()->GetClassLinker()->EnsureInitialized(self, h_class, true, true)) { in CheckClassInitializedForObjectAlloc() 137 DCHECK(self->IsExceptionPending()); in CheckClassInitializedForObjectAlloc() 140 DCHECK(!self->IsExceptionPending()); in CheckClassInitializedForObjectAlloc() 148 Thread* self, in CheckObjectAlloc() argument 153 self->ThrowNewException("Ljava/lang/InstantiationError;", klass->PrettyDescriptor().c_str()); in CheckObjectAlloc() 163 return CheckClassInitializedForObjectAlloc(klass, self, slow_path); in CheckObjectAlloc() 171 Thread* self, in AllocObjectFromCode() argument 174 klass = CheckObjectAlloc(klass, self, &slow_path); in AllocObjectFromCode() [all …]
|
/art/runtime/gc/allocator/ |
D | rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, in Alloc() argument 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, in Alloc() 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in Alloc() 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size, in Alloc() 59 inline bool RosAlloc::CanAllocFromThreadLocalRun(Thread* self, size_t size) { in CanAllocFromThreadLocalRun() argument 66 Run* thread_local_run = reinterpret_cast<Run*>(self->GetRosAllocRun(idx)); in CanAllocFromThreadLocalRun() 69 MutexLock mu(self, *size_bracket_locks_[idx]); in CanAllocFromThreadLocalRun() 78 inline void* RosAlloc::AllocFromThreadLocalRun(Thread* self, size_t size, in AllocFromThreadLocalRun() argument 86 Run* thread_local_run = reinterpret_cast<Run*>(self->GetRosAllocRun(idx)); in AllocFromThreadLocalRun() 89 MutexLock mu(self, *size_bracket_locks_[idx]); in AllocFromThreadLocalRun()
|
/art/runtime/mirror/ |
D | method_handles_lookup.cc | 32 ObjPtr<MethodHandlesLookup> MethodHandlesLookup::Create(Thread* const self, in Create() argument 38 GetClassRoot<MethodHandlesLookup>()->AllocObject(self)); in Create() 44 ObjPtr<MethodHandlesLookup> MethodHandlesLookup::GetDefault(Thread* const self) { in GetDefault() argument 47 lookup->Invoke(self, nullptr, 0, &result, "L"); in GetDefault() 51 ObjPtr<MethodHandle> MethodHandlesLookup::FindConstructor(Thread* const self, in FindConstructor() argument 62 findConstructor->Invoke(self, args, sizeof(args), &result, "LLL"); in FindConstructor()
|