/art/runtime/interpreter/mterp/arm/ |
D | object.S | 9 mov r3, rSELF @ arg3: Thread* self 28 mov r3, rSELF @ r3<- self 29 bl MterpCheckCast @ (index, &obj, method, self) 179 mov r3, rSELF @ r3<- self 180 bl MterpInstanceOf @ (index, &obj, method, self) 275 bl MterpNewInstance @ (shadow_frame, self, inst_data)
|
/art/runtime/ |
D | reference_table_test.cc | 47 Thread* self = Thread::Current(); in CreateWeakReference() local 50 StackHandleScope<3> scope(self); in CreateWeakReference() 54 class_linker->FindClass(self, in CreateWeakReference() 58 CHECK(class_linker->EnsureInitialized(self, h_ref_class, true, true)); in CreateWeakReference() 61 h_ref_class->AllocObject(self))); in CreateWeakReference() 72 constructor->Invoke(self, args, sizeof(uint32_t), &result, constructor->GetShorty()); in CreateWeakReference() 73 CHECK(!self->IsExceptionPending()); in CreateWeakReference()
|
D | quick_exception_handler.cc | 48 QuickExceptionHandler::QuickExceptionHandler(Thread* self, bool is_deoptimization) in QuickExceptionHandler() argument 49 : self_(self), in QuickExceptionHandler() 50 context_(self->GetLongJumpContext()), in QuickExceptionHandler() 66 CatchBlockStackVisitor(Thread* self, in CatchBlockStackVisitor() argument 72 : StackVisitor(self, context, StackVisitor::StackWalkKind::kIncludeInlinedFrames), in CatchBlockStackVisitor() 303 DeoptimizeStackVisitor(Thread* self, in DeoptimizeStackVisitor() argument 308 : StackVisitor(self, context, StackVisitor::StackWalkKind::kIncludeInlinedFrames), in DeoptimizeStackVisitor() 665 void QuickExceptionHandler::DumpFramesWithType(Thread* self, bool details) { in DumpFramesWithType() argument 697 self, in DumpFramesWithType()
|
D | class_linker-inl.h | 42 inline ObjPtr<mirror::Class> ClassLinker::FindArrayClass(Thread* self, in FindArrayClass() argument 56 ObjPtr<mirror::Class> array_class = FindClass(self, descriptor.c_str(), class_loader); in FindArrayClass() 64 self->AssertPendingException(); in FindArrayClass() 344 inline ArtMethod* ClassLinker::ResolveMethod(Thread* self, in ResolveMethod() argument 362 StackHandleScope<2> hs(self); in ResolveMethod() 455 Thread* const self = Thread::Current(); in VisitClassTables() local 456 WriterMutexLock mu(self, *Locks::classlinker_classes_lock_); in VisitClassTables() 473 void ClassLinker::VisitKnownDexFiles(Thread* self, Visitor visitor) { in VisitKnownDexFiles() argument 474 ReaderMutexLock rmu(self, *Locks::dex_lock_); in VisitKnownDexFiles()
|
D | debugger.h | 142 void ThreadStart(Thread* self) override REQUIRES_SHARED(Locks::mutator_lock_); 143 void ThreadDeath(Thread* self) override REQUIRES_SHARED(Locks::mutator_lock_);
|
D | cha.cc | 258 Thread* self = Thread::Current(); in Run() local 259 ScopedObjectAccess soa(self); in Run() 262 barrier_.Pass(self); in Run() 266 Thread* self = Thread::Current(); in WaitForThreadsToRunThroughCheckpoint() local 267 ScopedThreadStateChange tsc(self, kWaitingForCheckPointsToRun); in WaitForThreadsToRunThroughCheckpoint() 268 barrier_.Increment(self, threads_running_checkpoint); in WaitForThreadsToRunThroughCheckpoint() 634 Thread *self = Thread::Current(); in InvalidateSingleImplementationMethods() local 646 MutexLock cha_mu(self, *Locks::cha_lock_); in InvalidateSingleImplementationMethods()
|
D | thread.h | 275 bool ModifySuspendCount(Thread* self, 579 void Interrupt(Thread* self) REQUIRES(!wait_mutex_); 601 void NotifyLocked(Thread* self) REQUIRES(wait_mutex_); 763 return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values, self)); in SelfOffset() 1447 bool ModifySuspendCountInternal(Thread* self, 1460 bool PassActiveSuspendBarriers(Thread* self) 1665 self(nullptr), opeer(nullptr), jpeer(nullptr), stack_begin(nullptr), stack_size(0), in PACKED() 1708 Thread* self; in PACKED() local 1950 ScopedStackedShadowFramePusher(Thread* self, ShadowFrame* sf, StackedShadowFrameType type) in ScopedStackedShadowFramePusher() argument 1951 : self_(self), type_(type) { in ScopedStackedShadowFramePusher() [all …]
|
D | obj_ptr-inl.h | 30 Thread* self = Thread::Current(); in GetCurrentTrimedCookie() local 31 if (UNLIKELY(self == nullptr)) { in GetCurrentTrimedCookie() 34 return self->GetPoisonObjectCookie() & kCookieMask; in GetCurrentTrimedCookie()
|
D | signal_catcher.h | 51 int WaitForSignal(Thread* self, SignalSet& signals) REQUIRES(!lock_);
|
/art/openjdkjvmti/ |
D | ti_method.cc | 561 void Run(art::Thread* self) override REQUIRES_SHARED(art::Locks::mutator_lock_) { in Run() argument 567 FindFrameAtDepthVisitor visitor(self, context.get(), depth_); in Run() 607 DeoptManager::Get()->DeoptimizeThread(self); in Run() 634 art::Thread* self = art::Thread::Current(); in InferSlotTypeFromVerifier() local 635 art::StackHandleScope<2> hs(self); in InferSlotTypeFromVerifier() 638 self, in InferSlotTypeFromVerifier() 946 art::Thread* self = art::Thread::Current(); in GetLocalVariableGeneric() local 947 art::ScopedObjectAccess soa(self); in GetLocalVariableGeneric() 948 art::Locks::thread_list_lock_->ExclusiveLock(self); in GetLocalVariableGeneric() 952 art::Locks::thread_list_lock_->ExclusiveUnlock(self); in GetLocalVariableGeneric() [all …]
|
D | ti_heap.h | 95 art::Thread* self, 101 static void ReplaceReference(art::Thread* self,
|
/art/test/1908-suspend-native-resume-self/ |
D | expected.txt | 7 other thread attempting self resume
|
/art/runtime/gc/space/ |
D | bump_pointer_space-inl.h | 43 inline mirror::Object* BumpPointerSpace::AllocThreadUnsafe(Thread* self, size_t num_bytes, in AllocThreadUnsafe() argument 47 Locks::mutator_lock_->AssertExclusiveHeld(self); in AllocThreadUnsafe()
|
D | bump_pointer_space.h | 55 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 144 bool AllocNewTlab(Thread* self, size_t bytes) REQUIRES(!block_lock_);
|
D | malloc_space.cc | 262 Thread* self = context->self; in SweepCallback() local 263 Locks::heap_bitmap_lock_->AssertExclusiveHeld(self); in SweepCallback() 276 context->freed.bytes += space->FreeList(self, num_ptrs, ptrs); in SweepCallback()
|
/art/compiler/jit/ |
D | jit_compiler.cc | 168 Thread* self, JitMemoryRegion* region, ArtMethod* method, CompilationKind compilation_kind) { in CompileMethod() argument 178 self->AssertNoPendingException(); in CompileMethod() 193 self, code_cache, region, method, compilation_kind, jit_logger_.get()); in CompileMethod()
|
/art/dex2oat/linker/ |
D | image_writer.cc | 153 Thread* self, Runtime* runtime) REQUIRES_SHARED(Locks::mutator_lock_) { in AllocateBootImageLiveObjects() argument 158 StackHandleScope<1> hs(self); in AllocateBootImageLiveObjects() 160 hs.NewHandle(IntrinsicObjects::LookupIntegerCache(self, class_linker)); in AllocateBootImageLiveObjects() 166 self, GetClassRoot<mirror::ObjectArray<mirror::Object>>(class_linker), live_objects_size); in AllocateBootImageLiveObjects() 243 Thread* const self = Thread::Current(); in PrepareImageAddressSpace() local 247 ScopedObjectAccess soa(self); in PrepareImageAddressSpace() 268 ScopedObjectAccess soa(self); in PrepareImageAddressSpace() 276 ScopedObjectAccess soa(self); in PrepareImageAddressSpace() 283 ScopedObjectAccess soa(self); in PrepareImageAddressSpace() 285 std::vector<ObjPtr<mirror::DexCache>> dex_caches = FindDexCaches(self); in PrepareImageAddressSpace() [all …]
|
/art/runtime/verifier/ |
D | verifier_deps.h | 130 bool ValidateDependencies(Thread* self, 336 Thread* self, 365 Thread* self, 374 Thread* self, 384 Thread* self, 395 Thread* self,
|
/art/compiler/optimizing/ |
D | intrinsic_objects.cc | 31 Thread* self, ClassLinker* class_linker) { in LookupIntegerCache() argument 33 self, "Ljava/lang/Integer$IntegerCache;", /* class_loader= */ nullptr); in LookupIntegerCache()
|
/art/runtime/mirror/ |
D | dex_cache.cc | 39 void DexCache::InitializeDexCache(Thread* self, in InitializeDexCache() argument 58 ? reinterpret_cast<uint8_t*>(linear_alloc->AllocAlign16(self, layout.Size())) in InitializeDexCache() 59 : reinterpret_cast<uint8_t*>(linear_alloc->Alloc(self, layout.Size())); in InitializeDexCache() 220 Thread* const self = Thread::Current(); in AddPreResolvedStringsArray() local 225 linear_alloc->AllocArray<GcRoot<mirror::String>>(self, num_strings); in AddPreResolvedStringsArray()
|
D | array.h | 47 ALWAYS_INLINE static ObjPtr<Array> Alloc(Thread* self, 55 static ObjPtr<Array> CreateMultiArray(Thread* self, 128 static ObjPtr<Array> CopyOf(Handle<Array> h_this, Thread* self, int32_t new_length) 154 static ObjPtr<PrimitiveArray<T>> Alloc(Thread* self, size_t length) 157 static ObjPtr<PrimitiveArray<T>> AllocateAndFill(Thread* self, const T* data, size_t length)
|
D | class_ext-inl.h | 45 Thread* self = Thread::Current(); in EnsureJniIdsArrayPresent() local 46 StackHandleScope<2> hs(self); in EnsureJniIdsArrayPresent() 52 new_arr.Assign(Runtime::Current()->GetClassLinker()->AllocPointerArray(self, count)); in EnsureJniIdsArrayPresent() 56 self->AssertPendingOOMException(); in EnsureJniIdsArrayPresent()
|
/art/runtime/gc/collector/ |
D | mark_sweep.h | 90 void MarkRoots(Thread* self) 105 void MarkRootsCheckpoint(Thread* self, bool revoke_ros_alloc_thread_local_buffers_at_checkpoint) 132 void ProcessReferences(Thread* self) 176 void SweepSystemWeaks(Thread* self) 301 void RevokeAllThreadLocalAllocationStacks(Thread* self) NO_THREAD_SAFETY_ANALYSIS;
|
/art/runtime/arch/x86_64/ |
D | thread_x86_64.cc | 59 tlsPtr_.self = this; in InitCpu()
|
/art/compiler/driver/ |
D | compiled_method_storage.cc | 170 Thread* self = Thread::Current(); in DumpMemoryUsage() local 171 os << "\nCode dedupe: " << dedupe_code_.DumpStats(self); in DumpMemoryUsage() 172 os << "\nVmap table dedupe: " << dedupe_vmap_table_.DumpStats(self); in DumpMemoryUsage() 173 os << "\nCFI info dedupe: " << dedupe_cfi_info_.DumpStats(self); in DumpMemoryUsage()
|