Home
last modified time | relevance | path

Searched refs:self (Results 101 – 125 of 397) sorted by relevance

12345678910>>...16

/art/runtime/
Dnon_debuggable_classes.cc31 Thread* self = Thread::Current(); in AddNonDebuggableClass() local
32 JNIEnvExt* env = self->GetJniEnv(); in AddNonDebuggableClass()
33 ObjPtr<mirror::Class> mirror_klass(self->DecodeJObject(klass)->AsClass()); in AddNonDebuggableClass()
35 if (self->DecodeJObject(c)->AsClass() == mirror_klass) { in AddNonDebuggableClass()
Dmonitor.h88 static void Notify(Thread* self, ObjPtr<mirror::Object> obj) in Notify() argument
90 DoNotify(self, obj, false); in Notify()
92 static void NotifyAll(Thread* self, ObjPtr<mirror::Object> obj) in NotifyAll() argument
94 DoNotify(self, obj, true); in NotifyAll()
99 static void Wait(Thread* self,
152 static void InflateThinLocked(Thread* self, Handle<mirror::Object> obj, LockWord lock_word,
158 static bool Deflate(Thread* self, ObjPtr<mirror::Object> obj)
176 Monitor(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code)
178 Monitor(Thread* self, Thread* owner, ObjPtr<mirror::Object> obj, int32_t hash_code, MonitorId id)
183 bool Install(Thread* self)
[all …]
Dlinear_alloc.h32 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_);
33 void* AllocAlign16(Thread* self, size_t size) REQUIRES(!lock_);
36 void* Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) REQUIRES(!lock_);
40 T* AllocArray(Thread* self, size_t elements) REQUIRES(!lock_) { in AllocArray() argument
41 return reinterpret_cast<T*>(Alloc(self, elements * sizeof(T))); in AllocArray()
Dimtable_test.cc46 Thread* self = Thread::Current(); in LoadClasses() local
48 StackHandleScope<3> hs(self); in LoadClasses()
53 ObjPtr<mirror::ClassLoader>::DownCast(self->DecodeJObject(jclass_loader_a))); in LoadClasses()
55 hs.NewHandle(class_linker->FindClass(self, class_name.c_str(), h_class_loader))); in LoadClasses()
57 LOG(ERROR) << self->GetException()->Dump(); in LoadClasses()
63 ObjPtr<mirror::ClassLoader>::DownCast(self->DecodeJObject(jclass_loader_b))); in LoadClasses()
65 hs.NewHandle(class_linker->FindClass(self, class_name.c_str(), h_class_loader))); in LoadClasses()
67 LOG(ERROR) << self->GetException()->Dump(); in LoadClasses()
Dintern_table.cc92 ObjPtr<mirror::String> InternTable::LookupWeak(Thread* self, ObjPtr<mirror::String> s) { in LookupWeak() argument
93 MutexLock mu(self, *Locks::intern_table_lock_); in LookupWeak()
97 ObjPtr<mirror::String> InternTable::LookupStrong(Thread* self, ObjPtr<mirror::String> s) { in LookupStrong() argument
98 MutexLock mu(self, *Locks::intern_table_lock_); in LookupStrong()
102 ObjPtr<mirror::String> InternTable::LookupStrong(Thread* self, in LookupStrong() argument
109 MutexLock mu(self, *Locks::intern_table_lock_); in LookupStrong()
182 Thread* self = Thread::Current(); in BroadcastForNewInterns() local
183 MutexLock mu(self, *Locks::intern_table_lock_); in BroadcastForNewInterns()
184 weak_intern_condition_.Broadcast(self); in BroadcastForNewInterns()
187 void InternTable::WaitUntilAccessible(Thread* self) { in WaitUntilAccessible() argument
[all …]
Dvar_handles.cc30 bool VarHandleInvokeAccessorWithConversions(Thread* self, in VarHandleInvokeAccessorWithConversions() argument
38 StackHandleScope<1> hs(self); in VarHandleInvokeAccessorWithConversions()
40 var_handle->GetMethodTypeForAccessMode(self, access_mode))); in VarHandleInvokeAccessorWithConversions()
48 if (!PerformConversions(self, callsite_type, accessor_type, &getter, &setter, num_params)) { in VarHandleInvokeAccessorWithConversions()
61 bool VarHandleInvokeAccessor(Thread* self, in VarHandleInvokeAccessor() argument
83 return VarHandleInvokeAccessorWithConversions(self, in VarHandleInvokeAccessor()
/art/runtime/mirror/
Dmethod_type.cc31 ObjPtr<ObjectArray<Class>> AllocatePTypesArray(Thread* self, int count) in AllocatePTypesArray() argument
34 return ObjectArray<Class>::Alloc(self, class_array_type, count); in AllocatePTypesArray()
39 ObjPtr<MethodType> MethodType::Create(Thread* const self, in Create() argument
42 StackHandleScope<1> hs(self); in Create()
44 hs.NewHandle(ObjPtr<MethodType>::DownCast(GetClassRoot<MethodType>()->AllocObject(self)))); in Create()
58 ObjPtr<MethodType> MethodType::CloneWithoutLeadingParameter(Thread* const self, in CloneWithoutLeadingParameter() argument
60 StackHandleScope<3> hs(self); in CloneWithoutLeadingParameter()
64 Handle<ObjectArray<Class>> dst_ptypes = hs.NewHandle(AllocatePTypesArray(self, dst_ptypes_count)); in CloneWithoutLeadingParameter()
71 return Create(self, dst_rtype, dst_ptypes); in CloneWithoutLeadingParameter()
74 ObjPtr<MethodType> MethodType::CollectTrailingArguments(Thread* self, in CollectTrailingArguments() argument
[all …]
Dobject_array-alloc-inl.h36 inline ObjPtr<ObjectArray<T>> ObjectArray<T>::Alloc(Thread* self, in Alloc() argument
40 ObjPtr<Array> array = Array::Alloc(self, in Alloc()
54 inline ObjPtr<ObjectArray<T>> ObjectArray<T>::Alloc(Thread* self, in Alloc() argument
57 return Alloc(self, in Alloc()
65 Thread* self, in CopyOf() argument
72 ObjPtr<ObjectArray<T>> new_array = Alloc(self, h_this->GetClass(), new_length, allocator_type); in CopyOf()
Dfield.cc57 ObjPtr<mirror::Field> Field::CreateFromArtField(Thread* self, in CreateFromArtField() argument
60 StackHandleScope<2> hs(self); in CreateFromArtField()
65 DCHECK(self->IsExceptionPending()); in CreateFromArtField()
70 mirror::Throwable* exception = self->GetException(); in CreateFromArtField()
74 self->ClearException(); in CreateFromArtField()
77 auto ret = hs.NewHandle(ObjPtr<Field>::DownCast(GetClassRoot<Field>()->AllocObject(self))); in CreateFromArtField()
79 self->AssertPendingOOMException(); in CreateFromArtField()
Dclass_ext.cc64 bool ClassExt::ExtendObsoleteArrays(Handle<ClassExt> h_this, Thread* self, uint32_t increase) { in ExtendObsoleteArrays() argument
66 StackHandleScope<4> hs(self); in ExtendObsoleteArrays()
79 cl->AllocPointerArray(self, new_len))); in ExtendObsoleteArrays()
82 self->AssertPendingOOMException(); in ExtendObsoleteArrays()
86 ObjectArray<DexCache>::Alloc(self, in ExtendObsoleteArrays()
87 cl->FindClass(self, in ExtendObsoleteArrays()
93 self->AssertPendingOOMException(); in ExtendObsoleteArrays()
117 ObjPtr<ClassExt> ClassExt::Alloc(Thread* self) { in Alloc() argument
118 return ObjPtr<ClassExt>::DownCast(GetClassRoot<ClassExt>()->AllocObject(self)); in Alloc()
/art/runtime/gc/space/
Dlarge_object_space_test.cc38 Thread* const self = Thread::Current(); in LargeObjectTest() local
63 mirror::Object* obj = los->Alloc(self, request_size, &allocation_size, nullptr, in LargeObjectTest()
86 ASSERT_FALSE(los->IsZygoteLargeObject(self, obj)); in LargeObjectTest()
88 los->SetAllLargeObjectsAsZygoteObjects(self, /*set_mark_bit=*/ false); in LargeObjectTest()
91 ASSERT_TRUE(los->IsZygoteLargeObject(self, obj)); in LargeObjectTest()
115 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr, in LargeObjectTest()
131 void Run(Thread* self) override { in Run() argument
134 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr, in Run()
139 los_->Free(self, ptr); in Run()
163 Thread* self = Thread::Current(); in RaceTest() local
[all …]
Dbump_pointer_space.cc112 Thread* self = Thread::Current(); in RevokeAllThreadLocalBuffers() local
113 MutexLock mu(self, *Locks::runtime_shutdown_lock_); in RevokeAllThreadLocalBuffers()
114 MutexLock mu2(self, *Locks::thread_list_lock_); in RevokeAllThreadLocalBuffers()
132 Thread* self = Thread::Current(); in AssertAllThreadLocalBuffersAreRevoked() local
133 MutexLock mu(self, *Locks::runtime_shutdown_lock_); in AssertAllThreadLocalBuffersAreRevoked()
134 MutexLock mu2(self, *Locks::thread_list_lock_); in AssertAllThreadLocalBuffersAreRevoked()
173 Thread* self = Thread::Current(); in GetBytesAllocated() local
174 MutexLock mu(self, *Locks::runtime_shutdown_lock_); in GetBytesAllocated()
175 MutexLock mu2(self, *Locks::thread_list_lock_); in GetBytesAllocated()
191 Thread* self = Thread::Current(); in GetObjectsAllocated() local
[all …]
Drosalloc_space.h56 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated,
59 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, in Alloc() argument
61 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, in Alloc()
64 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, in AllocThreadUnsafe() argument
67 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, in AllocThreadUnsafe()
73 size_t Free(Thread* self, mirror::Object* ptr) override
75 size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) override
78 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, in AllocNonvirtual() argument
81 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, in AllocNonvirtual()
84 mirror::Object* AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, in AllocNonvirtualThreadUnsafe() argument
[all …]
/art/runtime/arch/arm/
Dquick_entrypoints_cc_arm.cc24 Thread* self, JValue* result, uint32_t, uint32_t*,
29 Thread* self, JValue* result, const char* shorty) { in quick_invoke_reg_setup() argument
93 art_quick_invoke_stub_internal(method, args, args_size, self, result, result_in_float, in quick_invoke_reg_setup()
100 Thread* self, JValue* result, const char* shorty) { in art_quick_invoke_stub() argument
101 quick_invoke_reg_setup<false>(method, args, args_size, self, result, shorty); in art_quick_invoke_stub()
107 uint32_t args_size, Thread* self, JValue* result, in art_quick_invoke_static_stub() argument
109 quick_invoke_reg_setup<true>(method, args, args_size, self, result, shorty); in art_quick_invoke_static_stub()
/art/runtime/entrypoints/quick/
Dquick_trampoline_entrypoints.cc589 Thread* self = Thread::Current(); in HandleDeoptimization() local
604 QuickExceptionHandler::DumpFramesWithType(self, true); in HandleDeoptimization()
610 self->PopDeoptimizationContext(/* out */ result, in HandleDeoptimization()
616 self->PushManagedStackFragment(fragment); in HandleDeoptimization()
631 EntireStackVisitor esv(self); in HandleDeoptimization()
638 self->SetException(pending_exception); in HandleDeoptimization()
640 interpreter::EnterInterpreterFromDeoptimize(self, in HandleDeoptimization()
647 extern "C" uint64_t artQuickToInterpreterBridge(ArtMethod* method, Thread* self, ArtMethod** sp) in artQuickToInterpreterBridge() argument
651 ScopedQuickEntrypointChecks sqec(self); in artQuickToInterpreterBridge()
659 ShadowFrame* deopt_frame = self->PopStackedShadowFrame( in artQuickToInterpreterBridge()
[all …]
/art/runtime/gc/
Dtask_processor.h57 void AddTask(Thread* self, HeapTask* task) REQUIRES(!lock_);
58 HeapTask* GetTask(Thread* self) REQUIRES(!lock_);
59 void Start(Thread* self) REQUIRES(!lock_);
62 void Stop(Thread* self) REQUIRES(!lock_);
63 void RunAllTasks(Thread* self) REQUIRES(!lock_);
65 void UpdateTargetRunTime(Thread* self, HeapTask* target_time, uint64_t new_target_time)
Dscoped_gc_critical_section.cc50 ScopedGCCriticalSection::ScopedGCCriticalSection(Thread* self, in ScopedGCCriticalSection() argument
53 : critical_section_(self, "ScopedGCCriticalSection") { in ScopedGCCriticalSection()
62 Thread* self, in ScopedInterruptibleGCCriticalSection() argument
64 CollectorType type) : self_(self) { in ScopedInterruptibleGCCriticalSection()
65 DCHECK(self != nullptr); in ScopedInterruptibleGCCriticalSection()
Dreference_processor.h60 void BroadcastForSlowPath(Thread* self);
62 ObjPtr<mirror::Object> GetReferent(Thread* self, ObjPtr<mirror::Reference> reference)
66 SelfDeletingTask* CollectClearedReferences(Thread* self) REQUIRES(!Locks::mutator_lock_);
85 void DisableSlowPath(Thread* self) REQUIRES(Locks::reference_processor_lock_)
90 void StartPreservingReferences(Thread* self) REQUIRES(!Locks::reference_processor_lock_);
91 void StopPreservingReferences(Thread* self) REQUIRES(!Locks::reference_processor_lock_);
93 void WaitUntilDoneProcessingReferences(Thread* self)
Dheap.cc875 void Heap::IncrementDisableMovingGC(Thread* self) { in IncrementDisableMovingGC() argument
878 ScopedThreadStateChange tsc(self, kWaitingForGcToComplete); in IncrementDisableMovingGC()
879 MutexLock mu(self, *gc_complete_lock_); in IncrementDisableMovingGC()
882 WaitForGcToCompleteLocked(kGcCauseDisableMovingGc, self); in IncrementDisableMovingGC()
886 void Heap::DecrementDisableMovingGC(Thread* self) { in DecrementDisableMovingGC() argument
887 MutexLock mu(self, *gc_complete_lock_); in DecrementDisableMovingGC()
892 void Heap::IncrementDisableThreadFlip(Thread* self) { in IncrementDisableThreadFlip() argument
895 bool is_nested = self->GetDisableThreadFlipCount() > 0; in IncrementDisableThreadFlip()
896 self->IncrementDisableThreadFlipCount(); in IncrementDisableThreadFlip()
902 ScopedThreadStateChange tsc(self, kWaitingForGcThreadFlip); in IncrementDisableThreadFlip()
[all …]
/art/openjdkjvmti/
Djvmti_weak_table.h84 art::Thread* self = art::Thread::Current(); in GetTag() local
85 art::MutexLock mu(self, allow_disallow_lock_); in GetTag()
86 Wait(self); in GetTag()
88 return GetTagLocked(self, obj, result); in GetTag()
93 art::Thread* self = art::Thread::Current(); in GetTagLocked() local
94 allow_disallow_lock_.AssertHeld(self); in GetTagLocked()
95 Wait(self); in GetTagLocked()
97 return GetTagLocked(self, obj, result); in GetTagLocked()
135 bool SetLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, T tag)
140 bool RemoveLocked(art::Thread* self, art::ObjPtr<art::mirror::Object> obj, /* out */ T* tag)
[all …]
Dti_stack.cc168 void Run(art::Thread* self) override REQUIRES_SHARED(art::Locks::mutator_lock_) { in Run()
172 auto visitor = MakeStackTraceVisitor(self, start_input, stop_input, frames_fn); in Run()
235 void Run(art::Thread* self) override REQUIRES_SHARED(art::Locks::mutator_lock_) { in Run()
240 auto visitor = MakeStackTraceVisitor(self, start_input, stop_input, frames_fn); in Run()
332 art::Thread* self = art::Thread::Current(); in Run() local
333 Work(thread, self); in Run()
334 barrier.Pass(self); in Run()
337 void Work(art::Thread* thread, art::Thread* self) in Work()
345 std::vector<jvmtiFrameInfo>* thread_frames = data->GetFrameStorageFor(self, thread); in Work()
372 art::Thread* self = art::Thread::Current(); in RunCheckpointAndWait() local
[all …]
/art/test/970-iface-super-resolution-gen/util-src/
Dgenerate_java.py40 def __init__(self, inner): argument
41 self.inner = inner
43 def get_name(self): argument
44 return self.inner.get_name()
46 def __str__(self): argument
48 for line in str(self.inner).splitlines(keepends = True):
/art/runtime/gc/accounting/
Dmod_union_table_test.cc51 Thread* self, space::ContinuousMemMapAllocSpace* space, size_t component_count) in AllocObjectArray() argument
53 auto* klass = GetObjectArrayClass(self, space); in AllocObjectArray()
57 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated)); in AllocObjectArray()
72 mirror::Class* GetObjectArrayClass(Thread* self, space::ContinuousMemMapAllocSpace* space) in GetObjectArrayClass() argument
84 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated, in GetObjectArrayClass()
179 Thread* const self = Thread::Current(); in RunTest() local
180 ScopedObjectAccess soa(self); in RunTest()
191 ScopedThreadSuspension sts(self, kSuspended); in RunTest()
199 auto* obj1 = AllocObjectArray(self, space, CardTable::kCardSize); in RunTest()
201 auto* obj2 = AllocObjectArray(self, space, CardTable::kCardSize); in RunTest()
[all …]
/art/runtime/gc/collector/
Dconcurrent_copying.cc124 Thread* self = Thread::Current(); in ConcurrentCopying() local
126 ReaderMutexLock mu(self, *Locks::heap_bitmap_lock_); in ConcurrentCopying()
134 MutexLock mu(self, mark_stack_lock_); in ConcurrentCopying()
158 Thread* const self = Thread::Current(); in MarkHeapReference() local
165 mirror::Object* to_ref = Mark(self, from_ref); in MarkHeapReference()
178 field->Assign(Mark(self, field->AsMirrorPtr())); in MarkHeapReference()
190 Thread* self = Thread::Current(); in RunPhases() local
191 thread_running_gc_ = self; in RunPhases()
192 Locks::mutator_lock_->AssertNotHeld(self); in RunPhases()
194 ReaderMutexLock mu(self, *Locks::mutator_lock_); in RunPhases()
[all …]
/art/runtime/jit/
Djit.cc294 Thread* self, in CompileMethod() argument
337 method_to_compile, self, compilation_kind, prejit, region)) { in CompileMethod()
344 bool success = jit_compiler_->CompileMethod(self, region, method_to_compile, compilation_kind); in CompileMethod()
345 code_cache_->DoneCompiling(method_to_compile, self, compilation_kind); in CompileMethod()
352 if (self->IsExceptionPending()) { in CompileMethod()
353 mirror::Throwable* exception = self->GetException(); in CompileMethod()
370 Thread* self = Thread::Current(); in DeleteThreadPool() local
382 pool->StopWorkers(self); in DeleteThreadPool()
383 pool->RemoveAllTasks(self); in DeleteThreadPool()
388 pool->Wait(self, false, false); in DeleteThreadPool()
[all …]

12345678910>>...16