/art/compiler/optimizing/ |
D | licm_test.cc | 133 parameter_, int_constant_, nullptr, DataType::Type::kInt32, MemberOffset(20), in TEST_F() 183 parameter_, int_constant_, DataType::Type::kInt32, 0); in TEST_F()
|
D | nodes_shared.cc | 50 if (result_type == DataType::Type::kInt32 && input_type == DataType::Type::kInt64) { in GetOpInfoFromInstruction()
|
D | code_generator_x86.cc | 181 DataType::Type::kInt32, in EmitNativeCode() 184 DataType::Type::kInt32); in EmitNativeCode() 433 DataType::Type::kInt32, in EmitNativeCode() 833 DataType::Type::kInt32, in EmitNativeCode() 1220 case DataType::Type::kInt32: in GetReturnLocation() 1250 case DataType::Type::kInt32: { in GetNextLocation() 1379 DataType::Type::kInt32, in Move64() 1382 DataType::Type::kInt32); in Move64() 1401 size_t elem_size = DataType::Size(DataType::Type::kInt32); in Move64() 1434 DataType::Type::kInt32, in Move64() [all …]
|
D | nodes_shared.h | 153 DataType::Type::kInt32, in HIntermediateAddressIndex() 207 shift_amount_(shift & (instr->GetType() == DataType::Type::kInt32
|
D | superblock_cloner_test.cc | 40 DataType::Type::kInt32)); in InitGraphAndParameters() 72 HPhi* phi = new (GetAllocator()) HPhi(GetAllocator(), 0, 0, DataType::Type::kInt32); in CreateBasicLoopDataFlow() 86 new (GetAllocator()) HArrayGet(null_check, bounds_check, DataType::Type::kInt32, dex_pc); in CreateBasicLoopDataFlow() 87 HInstruction* add = new (GetAllocator()) HAdd(DataType::Type::kInt32, array_get, const_1); in CreateBasicLoopDataFlow() 89 null_check, bounds_check, add, DataType::Type::kInt32, dex_pc); in CreateBasicLoopDataFlow() 90 HInstruction* induction_inc = new (GetAllocator()) HAdd(DataType::Type::kInt32, phi, const_1); in CreateBasicLoopDataFlow() 435 HInstruction* temp_add = new (GetAllocator()) HAdd(DataType::Type::kInt32, in TEST_F()
|
D | instruction_simplifier_arm.cc | 93 if (type != DataType::Type::kInt32 && type != DataType::Type::kInt64) { in TryMergeIntoShifterOperand() 124 shift_amount &= use->GetType() == DataType::Type::kInt32 in TryMergeIntoShifterOperand()
|
D | intrinsics_arm64.cc | 247 case DataType::Type::kInt32: in GenReverseBytes() 262 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerReverseBytes() 284 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenNumberOfLeadingZeros() 297 GenNumberOfLeadingZeros(invoke->GetLocations(), DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerNumberOfLeadingZeros() 311 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenNumberOfTrailingZeros() 325 GenNumberOfTrailingZeros(invoke->GetLocations(), DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerNumberOfTrailingZeros() 339 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenReverse() 352 GenReverse(invoke->GetLocations(), DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerReverse() 365 DCHECK_EQ(instr->GetType(), DataType::Type::kInt32); in GenBitCount() 393 GenBitCount(invoke, DataType::Type::kInt32, GetVIXLAssembler()); in VisitIntegerBitCount() [all …]
|
D | code_generator_x86_64.cc | 125 if (type_ == DataType::Type::kInt32) { in EmitNativeCode() 230 DataType::Type::kInt32, in EmitNativeCode() 233 DataType::Type::kInt32); in EmitNativeCode() 444 DataType::Type::kInt32, in EmitNativeCode() 851 DataType::Type::kInt32, in EmitNativeCode() 1071 case DataType::Type::kInt32: in GenerateStaticOrDirectCall() 1712 case DataType::Type::kInt32: in GenerateCompareTest() 2238 case DataType::Type::kInt32: in VisitCompare() 2273 case DataType::Type::kInt32: { in VisitCompare() 2413 case DataType::Type::kInt32: in VisitReturn() [all …]
|
D | optimizing_unit_test.h | 65 LiveInterval::MakeInterval(allocator, DataType::Type::kInt32, defined_by); 156 DataType::Type return_type = DataType::Type::kInt32,
|
D | code_generator_arm_vixl.cc | 281 case DataType::Type::kInt32: in GetLoadOperandType() 306 case DataType::Type::kInt32: in GetStoreOperandType() 488 DataType::Type::kInt32, in EmitNativeCode() 491 DataType::Type::kInt32); in EmitNativeCode() 698 DataType::Type::kInt32, in EmitNativeCode() 857 DataType::Type::kInt32, in EmitNativeCode() 2336 case DataType::Type::kInt32: { in GetNextLocation() 2416 case DataType::Type::kInt32: { in GetReturnLocation() 3563 case DataType::Type::kInt32: { in VisitNeg() 3590 case DataType::Type::kInt32: in VisitNeg() [all …]
|
D | intrinsics_x86.cc | 280 case DataType::Type::kInt32: in GenReverseBytes() 294 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt32, GetAssembler()); in VisitIntegerReverseBytes() 1484 case DataType::Type::kInt32: in GenPeek() 1510 GenPeek(invoke->GetLocations(), DataType::Type::kInt32, GetAssembler()); in VisitMemoryPeekIntNative() 1565 case DataType::Type::kInt32: in GenPoke() 1598 CreateLongIntToVoidLocations(allocator_, DataType::Type::kInt32, invoke); in VisitMemoryPokeIntNative() 1602 GenPoke(invoke->GetLocations(), DataType::Type::kInt32, GetAssembler()); in VisitMemoryPokeIntNative() 1645 case DataType::Type::kInt32: { in GenUnsafeGet() 1728 allocator_, invoke, DataType::Type::kInt32, /* is_volatile= */ false); in VisitUnsafeGet() 1731 …CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt32, /* is_volatile= */ true); in VisitUnsafeGetVolatile() [all …]
|
D | intrinsics_x86_64.cc | 198 case DataType::Type::kInt32: in GenReverseBytes() 215 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt32, GetAssembler()); in VisitIntegerReverseBytes() 1647 case DataType::Type::kInt32: in GenPeek() 1672 GenPeek(invoke->GetLocations(), DataType::Type::kInt32, GetAssembler()); in VisitMemoryPeekIntNative() 1720 case DataType::Type::kInt32: in GenPoke() 1757 GenPoke(invoke->GetLocations(), DataType::Type::kInt32, GetAssembler()); in VisitMemoryPokeIntNative() 1802 case DataType::Type::kInt32: in GenUnsafeGet() 1875 GenUnsafeGet(invoke, DataType::Type::kInt32, /* is_volatile= */ false, codegen_); in VisitUnsafeGet() 1878 GenUnsafeGet(invoke, DataType::Type::kInt32, /* is_volatile= */ true, codegen_); in VisitUnsafeGetVolatile() 1911 CreateIntIntIntIntToVoidPlusTempsLocations(allocator_, DataType::Type::kInt32, invoke); in VisitUnsafePut() [all …]
|
D | code_generator_arm64.cc | 266 DataType::Type::kInt32, in EmitNativeCode() 269 DataType::Type::kInt32); in EmitNativeCode() 559 DataType::Type::kInt32, in EmitNativeCode() 665 Register index_reg = RegisterFrom(index_, DataType::Type::kInt32); in EmitNativeCode() 739 DataType::Type::kInt32, in EmitNativeCode() 1308 __ Mov(RegisterFrom(location, DataType::Type::kInt32), value); in MoveConstant() 1442 return (cst->IsIntConstant() && type == DataType::Type::kInt32) || in CoherentConstantAndType() 1444 (cst->IsNullConstant() && type == DataType::Type::kInt32) || in CoherentConstantAndType() 1484 dst_type = destination.IsRegister() ? DataType::Type::kInt32 : DataType::Type::kFloat32; in MoveLocation() 1511 : DataType::Type::kInt32; in MoveLocation() [all …]
|
D | intrinsics_arm_vixl.cc | 300 DCHECK((type == DataType::Type::kInt32) || (type == DataType::Type::kInt64)); in GenNumberOfLeadingZeros() 324 GenNumberOfLeadingZeros(invoke, DataType::Type::kInt32, codegen_); in VisitIntegerNumberOfLeadingZeros() 338 DCHECK((type == DataType::Type::kInt32) || (type == DataType::Type::kInt64)); in GenNumberOfTrailingZeros() 370 GenNumberOfTrailingZeros(invoke, DataType::Type::kInt32, codegen_); in VisitIntegerNumberOfTrailingZeros() 577 case DataType::Type::kInt32: { in GenUnsafeGet() 667 CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt32); in VisitUnsafeGet() 670 CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt32); in VisitUnsafeGetVolatile() 686 GenUnsafeGet(invoke, DataType::Type::kInt32, /* is_volatile= */ false, codegen_); in VisitUnsafeGet() 689 GenUnsafeGet(invoke, DataType::Type::kInt32, /* is_volatile= */ true, codegen_); in VisitUnsafeGetVolatile() 731 allocator_, features_, DataType::Type::kInt32, /* is_volatile= */ false, invoke); in VisitUnsafePut() [all …]
|
D | ssa_builder.cc | 45 (right->GetType() == DataType::Type::kInt32)) { in FixNullConstantType() 48 (left->GetType() == DataType::Type::kInt32)) { in FixNullConstantType() 302 type == DataType::Type::kInt32 ? DataType::Type::kFloat32 : DataType::Type::kFloat64, in CreateFloatOrDoubleEquivalentOfArrayGet()
|
D | register_allocation_resolver.cc | 115 case DataType::Type::kInt32: in Resolve() 211 case DataType::Type::kInt32: in Resolve() 517 move->AddMove(source.ToLow(), destination.ToLow(), DataType::Type::kInt32, instruction); in AddMove() 518 move->AddMove(source.ToHigh(), destination.ToHigh(), DataType::Type::kInt32, nullptr); in AddMove()
|
D | bounds_check_elimination.cc | 952 && (phi->GetType() == DataType::Type::kInt32) in VisitPhi() 1027 if (instruction->GetResultType() != DataType::Type::kInt32) { in TryToAssignRangeIfOptimizedRemWithConstantDivisor() 1081 if (const_divisor == DataType::MinValueOfIntegralType(DataType::Type::kInt32)) { in TryToAssignRangeIfOptimizedRemWithConstantDivisor() 1083 ValueBound(nullptr, DataType::MinValueOfIntegralType(DataType::Type::kInt32) + 1), in TryToAssignRangeIfOptimizedRemWithConstantDivisor() 1084 ValueBound(nullptr, DataType::MaxValueOfIntegralType(DataType::Type::kInt32))); in TryToAssignRangeIfOptimizedRemWithConstantDivisor() 1086 DCHECK_GT(const_divisor, DataType::MinValueOfIntegralType(DataType::Type::kInt32)); in TryToAssignRangeIfOptimizedRemWithConstantDivisor() 1087 DCHECK_LE(const_divisor, DataType::MaxValueOfIntegralType(DataType::Type::kInt32)); in TryToAssignRangeIfOptimizedRemWithConstantDivisor() 1386 HAdd(DataType::Type::kInt32, base, GetGraph()->GetIntConstant(min_c)); in AddCompareWithDeoptimization() 1387 upper = new (GetGraph()->GetAllocator()) HAdd(DataType::Type::kInt32, base, upper); in AddCompareWithDeoptimization()
|
D | loop_optimization.cc | 135 case DataType::Type::kInt32: in IsSignExtensionAndGet() 139 from == DataType::Type::kInt32)) { in IsSignExtensionAndGet() 200 case DataType::Type::kInt32: in IsZeroExtensionAndGet() 1067 DCHECK(induc_type == DataType::Type::kInt32 || induc_type == DataType::Type::kInt64) in Vectorize() 1429 } else if (to == DataType::Type::kFloat32 && from == DataType::Type::kInt32) { in VectorizeUse() 1570 case DataType::Type::kInt32: in TrySetVectorType() 1590 case DataType::Type::kInt32: in TrySetVectorType() 1637 case DataType::Type::kInt32: in TrySetVectorType() 1706 subscript = new (global_allocator_) HAdd(DataType::Type::kInt32, subscript, offset); in GenerateVecSub() 2043 (reduction_type != DataType::Type::kInt32 && reduction_type != DataType::Type::kInt64)) { in VectorizeSADIdiom() [all …]
|
D | nodes_x86.h | 28 DataType::Type::kInt32, in HX86ComputeBaseMethodAddress()
|
D | codegen_test.cc | 628 if (type == DataType::Type::kInt32) { in TestComparison() 698 static_cast<IfCondition>(cond), i, j, DataType::Type::kInt32, target_config); in TEST_F() 732 move->AddMove(Location::StackSlot(0), Location::StackSlot(8192), DataType::Type::kInt32, nullptr); in TEST_F() 733 move->AddMove(Location::StackSlot(8192), Location::StackSlot(0), DataType::Type::kInt32, nullptr); in TEST_F()
|
D | constant_folding_test.cc | 42 DataType::Type return_type = DataType::Type::kInt32) { in TestCode() 750 graph_->GetDexFile(), dex::TypeIndex(0), 0, DataType::Type::kInt32, true); in TEST_F()
|
D | instruction_simplifier.cc | 281 type == DataType::Type::kInt32)) { in TryCombineVecMultiplyAccumulate() 370 DCHECK_EQ(shift_amount->GetType(), DataType::Type::kInt32); in VisitShift() 851 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in NewIntegralAbs() 864 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in NewIntegralMinMax() 890 (type1 == DataType::Type::kInt32 && type2 == DataType::Type::kInt32 && in AreLowerPrecisionArgs() 1395 HConstant* new_const = GetGraph()->GetConstant(DataType::Type::kInt32, value); in VisitAnd() 1397 new (GetGraph()->GetAllocator()) HAnd(DataType::Type::kInt32, new_and_input, new_const); in VisitAnd() 2290 case DataType::Type::kInt32: in SimplifySystemArrayCopy() 2832 if (type == DataType::Type::kInt32) { in ComputeAddition()
|
D | induction_var_analysis.cc | 66 to == DataType::Type::kInt32; in IsNarrowingIntegralConversion() 67 case DataType::Type::kInt32: in IsNarrowingIntegralConversion() 90 return DataType::Type::kInt32; in ImplicitConversion() 945 } else if (type != DataType::Type::kInt32 && type != DataType::Type::kInt64) { in VisitCondition() 1316 if (type == DataType::Type::kInt32 && 0 <= value && value < 31) { in GetShiftConstant() 1361 (info->type == DataType::Type::kInt32 && (info->op_a->type == DataType::Type::kInt64 || in IsNarrowingLinear()
|
D | instruction_simplifier_arm64.cc | 99 if (type != DataType::Type::kInt32 && type != DataType::Type::kInt64) { in TryMergeIntoShifterOperand()
|
D | instruction_simplifier_shared.cc | 95 if (type != DataType::Type::kInt32) { in TryCombineMultiplyAccumulate()
|