/art/compiler/optimizing/ |
D | graph_checker.cc | 64 size_t current_size = GetGraph()->GetReversePostOrder().size(); in Run() 90 ScopedArenaAllocator allocator(GetGraph()->GetArenaStack()); in VisitBasicBlock() 142 if (GetGraph()->IsInSsaForm() && !block->EndsWithControlFlowInstruction()) { in VisitBasicBlock() 294 if (!GetGraph()->HasBoundsChecks()) { in VisitBoundsCheck() 306 if (GetGraph()->IsCompilingOsr()) { in VisitDeoptimize() 758 HBasicBlock* loop_block = GetGraph()->GetBlocks()[i]; in HandleLoop() 930 ScopedArenaAllocator allocator(GetGraph()->GetArenaStack()); in VisitPhi() 934 GetGraph()->GetCurrentInstructionId(), in VisitPhi()
|
D | locations.cc | 56 instruction->GetBlock()->GetGraph()->GetAllocator()) {} in LocationSummary()
|
D | register_allocator.cc | 60 for (HBasicBlock* block : codegen_->GetGraph()->GetLinearOrder()) { in ~RegisterAllocator() 107 ScopedArenaAllocator allocator(codegen.GetGraph()->GetArenaStack()); in ValidateIntervals()
|
D | reference_type_propagation.cc | 103 return GetGraph()->GetHandleCache(); in GetHandleCache() 241 bound_type = new (receiver->GetBlock()->GetGraph()->GetAllocator()) HBoundType(receiver); in BoundTypeIn() 541 instr->SetReferenceTypeInfo(GetGraph()->GetInexactObjectRti()); in SetClassAsTypeInfo() 611 instr->SetReferenceTypeInfo(GetGraph()->GetInexactObjectRti()); in VisitUnresolvedInstanceFieldGet() 619 instr->SetReferenceTypeInfo(GetGraph()->GetInexactObjectRti()); in VisitUnresolvedStaticFieldGet() 840 instr->SetReferenceTypeInfo(GetGraph()->GetInexactObjectRti()); in UpdateArrayGet() 927 instr->SetReferenceTypeInfo(instr->GetBlock()->GetGraph()->GetInexactObjectRti()); in UpdatePhi()
|
D | code_generator_arm_vixl.cc | 529 DCHECK(IsSameDexFile(cls_->GetDexFile(), arm_codegen->GetGraph()->GetDexFile())); in EmitNativeCode() 689 HParallelMove parallel_move(codegen->GetGraph()->GetAllocator()); in EmitNativeCode() 845 HParallelMove parallel_move(codegen->GetGraph()->GetAllocator()); in EmitNativeCode() 2046 if (GetGraph()->IsDebuggable()) { in SetupBlockedRegisters() 2106 if (GetGraph()->IsCompilingBaseline() && !Runtime::Current()->IsAotCompiler()) { in MaybeIncrementHotness() 2108 ProfilingInfo* info = GetGraph()->GetArtMethod()->GetProfilingInfo(kRuntimePointerSize); in MaybeIncrementHotness() 2257 if (GetGraph()->HasShouldDeoptimizeFlag()) { in GenerateFrameEntry() 2542 HParallelMove move(GetGraph()->GetAllocator()); in MoveLocation() 2782 LocationSummary* locations = new (GetGraph()->GetAllocator()) LocationSummary(if_instr); in VisitIf() 2799 LocationSummary* locations = new (GetGraph()->GetAllocator()) in VisitDeoptimize() [all …]
|
D | code_generator.h | 195 HGraph* GetGraph() const { return graph_; } in GetGraph() function 334 DCHECK(GetGraph()->HasShouldDeoptimizeFlag()); in GetStackOffsetOfShouldDeoptimizeFlag() 580 HParallelMove parallel_move(GetGraph()->GetAllocator()); in PrepareCriticalNativeCall() 761 size_t size = GetGraph()->GetBlocks().size(); in CommonInitializeLabels() 763 GetGraph()->GetAllocator()->AllocArray<LabelType>(size, kArenaAllocCodeGenerator); in CommonInitializeLabels()
|
D | dead_code_elimination.cc | 127 return condition->GetBlock()->GetGraph()->GetIntConstant( in Evaluate() 194 bound = new (obj->GetBlock()->GetGraph()->GetAllocator()) HBoundType(obj); in RemoveNonNullControlDependences()
|
D | gvn.cc | 320 if (!pure || instruction->GetBlock()->GetGraph()->HasIrreducibleLoops()) { in HashCode() 556 HBasicBlock* current_block = block->GetGraph()->GetBlocks()[block_id]; in FindVisitedBlockWithRecyclableSet()
|
D | graph_visualizer.cc | 467 ? GetGraph()->GetDexFile().PrettyMethod(invoke->GetDexMethodIndex(), kWithSignature) in VisitInvoke() 778 GetGraph()->GetEntryBlock()->GetBlockId()); in DumpDisassemblyBlockForFrameEntry() 798 GetGraph()->HasExitBlock() ? GetGraph()->GetExitBlock()->GetBlockId() : -1, in DumpDisassemblyBlockForSlowPaths()
|
D | instruction_simplifier_arm.cc | 140 new (GetGraph()->GetAllocator()) HDataProcWithShifterOp(use, in TryMergeIntoShifterOperand()
|
D | superblock_cloner.cc | 1018 HGraph* graph = orig_block->GetGraph(); in CloneBasicBlock() 1149 HGraph* graph = loop_header->GetGraph(); in DoLoopTransformationImpl() 1201 info->GetHeader()->GetGraph()->GetAllocator()->Adapter(kArenaAllocSuperblockCloner)), in LoopClonerSimpleHelper() 1203 info->GetHeader()->GetGraph()->GetAllocator()->Adapter(kArenaAllocSuperblockCloner)), in LoopClonerSimpleHelper()
|
D | ssa_liveness_analysis_test.cc | 46 HGraph* graph = block->GetGraph(); in CreateSuccessor()
|
D | intrinsics.h | 88 HParallelMove parallel_move(codegen->GetGraph()->GetAllocator()); in INTRINSICS_LIST()
|
D | superblock_cloner.h | 389 cloner_(info->GetHeader()->GetGraph(), &info->GetBlocks(), bb_map, hir_map, induction_range) { in LoopClonerHelper()
|
D | scheduler.cc | 578 ScopedArenaAllocator allocator(block->GetGraph()->GetArenaStack()); in Schedule() 627 HGraph* graph = block->GetGraph(); in Schedule()
|
D | code_sinking.cc | 48 if (instruction->GetBlock() == instruction->GetBlock()->GetGraph()->GetEntryBlock()) { in IsInterestingInstruction()
|
D | register_allocator_linear_scan.cc | 81 reserved_out_slots_ = ptr_size / kVRegSize + codegen->GetGraph()->GetMaximumNumberOfOutVRegs(); in RegisterAllocatorLinearScan() 114 for (HBasicBlock* block : codegen_->GetGraph()->GetLinearOrder()) { in AllocateRegisters() 168 for (HBasicBlock* block : codegen_->GetGraph()->GetLinearPostOrder()) { in AllocateRegistersInternal()
|
D | code_generator_arm64.h | 557 return GetGraph()->HasSIMD() in GetSlowPathFPWidth() 621 block_labels_.resize(GetGraph()->GetBlocks().size()); in Initialize() 629 jump_tables_.emplace_back(new (GetGraph()->GetAllocator()) JumpTableARM64(switch_instr)); in CreateJumpTable()
|
D | optimizing_compiler.cc | 690 ArenaVector<linker::LinkerPatch> linker_patches(codegen->GetGraph()->GetAllocator()->Adapter()); in EmitAndSortLinkerPatches() 1405 codegen->GetGraph()->HasShouldDeoptimizeFlag(), in JitCompile() 1406 codegen->GetGraph()->GetCHASingleImplementationList())) { in JitCompile()
|
D | intrinsics.cc | 245 ArenaAllocator* allocator = invoke->GetBlock()->GetGraph()->GetAllocator(); in ComputeIntegerValueOfLocations()
|
D | code_generator_arm_vixl.h | 554 block_labels_.resize(GetGraph()->GetBlocks().size()); in Initialize() 768 jump_tables_.emplace_back(new (GetGraph()->GetAllocator()) JumpTableARMVIXL(switch_instr)); in CreateJumpTable()
|
D | ssa_liveness_analysis.h | 1050 if (block_at_use.GetGraph()->HasIrreducibleLoops()) { in AddBackEdgeUses() 1286 HGraph* graph = instruction->GetBlock()->GetGraph(); in ShouldBeLiveForEnvironment()
|
D | ssa_builder.cc | 299 HArrayGet* equivalent = new (aget->GetBlock()->GetGraph()->GetAllocator()) HArrayGet( in CreateFloatOrDoubleEquivalentOfArrayGet()
|
D | register_allocator_graph_color.cc | 83 if (block->Dominates(block->GetGraph()->GetExitBlock())) { in CostForMoveAt() 593 reserved_out_slots_(codegen->GetGraph()->GetMaximumNumberOfOutVRegs()) { in RegisterAllocatorGraphColor() 784 for (HBasicBlock* block : codegen_->GetGraph()->GetLinearPostOrder()) { in ProcessInstructions()
|
D | code_generator_x86.h | 384 return GetGraph()->HasSIMD() in GetSlowPathFPWidth()
|