/art/tools/checker/match/ |
D | file.py | 159 def findMatchingLine(statement, c1Pass, scope, variables, excludeLines=[]): argument 170 for i in range(scope.start, scope.end): 175 raise MatchFailedException(statement, scope.start, variables) 189 assert self.cursor <= match.scope.end 192 self.handleNotQueue(MatchScope(self.cursor, match.scope.start)) 194 self.cursor = match.scope.end + 1 197 def handleDagQueue(self, scope): argument 216 match = findMatchingLine(statement, self.c1Pass, scope, variables, matchedLines) 218 assert match.scope.start == match.scope.end 219 assert match.scope.start not in matchedLines [all …]
|
/art/benchmark/jobject-benchmark/ |
D | info.txt | 7 Decoding local, weak, global, handle scope jobjects.
|
/art/test/1985-structural-redefine-stack-scope/ |
D | Android.bp | 3 name: "art-run-test-1985-structural-redefine-stack-scope",
|
/art/test/1987-structural-redefine-recursive-stack-scope/ |
D | Android.bp | 3 name: "art-run-test-1987-structural-redefine-recursive-stack-scope",
|
/art/test/1986-structural-redefine-multi-thread-stack-scope/ |
D | Android.bp | 3 name: "art-run-test-1986-structural-redefine-multi-thread-stack-scope",
|
/art/tools/art_verifier/ |
D | art_verifier.cc | 217 StackHandleScope<2> scope(soa.Self()); in ExecuteWithRuntime() local 218 Handle<mirror::ClassLoader> h_loader = scope.NewHandle( in ExecuteWithRuntime() 220 MutableHandle<mirror::Class> h_klass(scope.NewHandle<mirror::Class>(nullptr)); in ExecuteWithRuntime()
|
/art/runtime/ |
D | reference_table_test.cc | 50 StackHandleScope<3> scope(self); in CreateWeakReference() local 51 Handle<mirror::Object> h_referent(scope.NewHandle<mirror::Object>(referent)); in CreateWeakReference() 53 Handle<mirror::Class> h_ref_class(scope.NewHandle<mirror::Class>( in CreateWeakReference() 60 Handle<mirror::Object> h_ref_instance(scope.NewHandle<mirror::Object>( in CreateWeakReference()
|
D | common_throws.h | 200 void ThrowNoSuchFieldError(std::string_view scope,
|
D | thread.h | 924 void PushReflectiveHandleScope(BaseReflectiveHandleScope* scope) { in PushReflectiveHandleScope() argument 925 DCHECK_EQ(scope->GetLink(), tlsPtr_.top_reflective_handle_scope); in PushReflectiveHandleScope() 926 DCHECK_EQ(scope->GetThread(), this); in PushReflectiveHandleScope() 927 tlsPtr_.top_reflective_handle_scope = scope; in PushReflectiveHandleScope()
|
D | common_throws.cc | 391 void ThrowNoSuchFieldError(std::string_view scope, in ThrowNoSuchFieldError() argument 397 msg << "No " << scope << "field " << name << " of type " << type in ThrowNoSuchFieldError()
|
/art/compiler/optimizing/ |
D | code_generator_arm64.h | 308 virtual Location AllocateSIMDScratchLocation(vixl::aarch64::UseScratchRegisterScope* scope) = 0; 310 vixl::aarch64::UseScratchRegisterScope* scope) = 0; 452 Location AllocateSIMDScratchLocation(vixl::aarch64::UseScratchRegisterScope* scope) override; 454 vixl::aarch64::UseScratchRegisterScope* scope) override; 485 Location AllocateSIMDScratchLocation(vixl::aarch64::UseScratchRegisterScope* scope) override; 487 vixl::aarch64::UseScratchRegisterScope* scope) override;
|
D | optimizing_compiler.cc | 331 PassScope scope(optimizations[i]->GetPassName(), pass_observer); in RunOptimizations() local 337 scope.SetPassNotChanged(); in RunOptimizations() 571 PassScope scope(PrepareForRegisterAllocation::kPrepareForRegisterAllocationPassName, in AllocateRegisters() local 580 PassScope scope(SsaLivenessAnalysis::kLivenessPassName, pass_observer); in AllocateRegisters() local 584 PassScope scope(RegisterAllocator::kRegisterAllocatorPassName, pass_observer); in AllocateRegisters() local 828 PassScope scope(HGraphBuilder::kBuilderPassName, &pass_observer); in TryCompile() local 953 PassScope scope(HGraphBuilder::kBuilderPassName, &pass_observer); in TryCompileIntrinsic() local
|
D | code_generator_vector_arm64_sve.cc | 1516 vixl::aarch64::UseScratchRegisterScope* scope) { in AllocateSIMDScratchLocation() argument 1518 return LocationFrom(scope->AcquireVRegisterOfSize(kQRegSize)); in AllocateSIMDScratchLocation() 1522 vixl::aarch64::UseScratchRegisterScope* scope) { in FreeSIMDScratchLocation() argument 1524 scope->Release(QRegisterFrom(loc)); in FreeSIMDScratchLocation()
|
D | code_generator_vector_arm64_neon.cc | 1516 vixl::aarch64::UseScratchRegisterScope* scope) { in AllocateSIMDScratchLocation() argument 1518 return LocationFrom(scope->AcquireVRegisterOfSize(kQRegSize)); in AllocateSIMDScratchLocation() 1522 vixl::aarch64::UseScratchRegisterScope* scope) { in FreeSIMDScratchLocation() argument 1524 scope->Release(QRegisterFrom(loc)); in FreeSIMDScratchLocation()
|
D | code_generator_arm64.cc | 587 EmissionCheckScope scope(codegen->GetVIXLAssembler(), in EmitTable() local 2430 ExactAssemblyScope scope(masm, kInstructionSize, CodeBufferCheckScope::kExactSize); in VisitMultiplyAccumulate() local
|
/art/tools/jfuzz/ |
D | README.md | 110 finite in size and scope. Tests typically focus on validating particular 128 The randomness of fuzz testing implies that the size and scope of testing is no
|
/art/test/utils/ |
D | regen-test-files | 135 art-run-test-1985-structural-redefine-stack-scope 136 art-run-test-1986-structural-redefine-multi-thread-stack-scope 137 art-run-test-1987-structural-redefine-recursive-stack-scope
|
/art/tools/checker/ |
D | README | 31 therefore create a scope within which the statement is verified. 47 only valid within the scope of the defining group. Within a group they cannot 129 (of any depth, since nested branching is allowed), that variable will become global within the scope
|
/art/oatdump/ |
D | oatdump.cc | 2879 StackHandleScope<1> scope(soa.Self()); in DumpOatWithRuntime() local 2880 Handle<mirror::ClassLoader> loader_handle = scope.NewHandle( in DumpOatWithRuntime() 2982 StackHandleScope<1> scope(self); in Dump() local 2983 MutableHandle<mirror::ClassLoader> class_loader = scope.NewHandle<mirror::ClassLoader>(nullptr); in Dump() 3064 StackHandleScope<1> scope(self); in DumpImtStats() local 3065 MutableHandle<mirror::Class> h_klass(scope.NewHandle<mirror::Class>(nullptr)); in DumpImtStats() 3208 StackHandleScope<1> scope(Thread::Current()); in PrepareAndGetImTable() local 3209 Handle<mirror::Class> h_klass = scope.NewHandle<mirror::Class>(klass); in PrepareAndGetImTable()
|
/art/runtime/gc/ |
D | heap.h | 399 void GetInstances(VariableSizedHandleScope& scope, 408 void GetReferringObjects(VariableSizedHandleScope& scope,
|
D | heap.cc | 1987 void Heap::GetInstances(VariableSizedHandleScope& scope, in GetInstances() argument 1996 instances.push_back(scope.NewHandle(obj)); in GetInstances() 2003 void Heap::GetReferringObjects(VariableSizedHandleScope& scope, in GetReferringObjects() argument 2041 ReferringObjectsFinder finder(scope, o, max_count, referring_objects); in GetReferringObjects()
|
/art/runtime/entrypoints/quick/ |
D | quick_trampoline_entrypoints.cc | 2013 void Reset(uintptr_t* gpr_regs, uint32_t* fpr_regs, uintptr_t* stack_args, HandleScope* scope) { in Reset() argument 2015 handle_scope_ = scope; in Reset()
|
/art/test/ |
D | Android.bp | 640 "1985-structural-redefine-stack-scope/stack_scope.cc",
|