/art/compiler/optimizing/ |
D | data_type.h | 40 kInt64, enumerator 65 case Type::kInt64: in SizeShift() 92 case Type::kInt64: in Size() 119 case Type::kInt64: in IsIntegralType() 127 return type == Type::kInt32 || type == Type::kInt64; in IsIntOrLongType() 131 return type == Type::kUint64 || type == Type::kInt64 || type == Type::kFloat64; in Is64BitType() 151 case Type::kInt64: in Kind() 152 return Type::kInt64; in Kind() 176 case Type::kInt64: in MinValueOfIntegralType() 202 case Type::kInt64: in MaxValueOfIntegralType() [all …]
|
D | parallel_move_test.cc | 294 DataType::Type::kInt64, in TYPED_TEST() 306 DataType::Type::kInt64, in TYPED_TEST() 323 DataType::Type::kInt64, in TYPED_TEST() 353 DataType::Type::kInt64, in TYPED_TEST() 374 DataType::Type::kInt64, in TYPED_TEST() 395 DataType::Type::kInt64, in TYPED_TEST() 420 DataType::Type::kInt64, in TYPED_TEST() 425 DataType::Type::kInt64, in TYPED_TEST() 440 DataType::Type::kInt64, in TYPED_TEST() 445 DataType::Type::kInt64, in TYPED_TEST() [all …]
|
D | scheduler_arm.cc | 34 case DataType::Type::kInt64: in HandleBinaryOperationLantencies() 61 case DataType::Type::kInt64: in VisitMul() 77 case DataType::Type::kInt64: in HandleBitwiseOperationLantencies() 108 case DataType::Type::kInt64: { in VisitRor() 142 case DataType::Type::kInt64: in HandleShiftLatencies() 207 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in HandleGenerateLongTestConstant() 273 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in HandleGenerateLongTest() 306 if (type == DataType::Type::kInt64) { in HandleGenerateTest() 320 if (condition->GetLeft()->GetType() == DataType::Type::kInt64) { in CanGenerateTest() 358 DCHECK_EQ(cond->GetLeft()->GetType(), DataType::Type::kInt64); in HandleGenerateEqualLong() [all …]
|
D | data_type-inl.h | 34 case Primitive::kPrimLong: return DataType::Type::kInt64; in DataTypeFromPrimitive() 59 case DataType::Type::kInt64: return 'j'; // Java long (J). in TypeId()
|
D | instruction_simplifier_arm.cc | 93 if (type != DataType::Type::kInt32 && type != DataType::Type::kInt64) { in TryMergeIntoShifterOperand() 129 if (!use->IsAdd() && (!use->IsSub() || use->GetType() != DataType::Type::kInt64)) { in TryMergeIntoShifterOperand() 134 } else if (type == DataType::Type::kInt64 && shift_amount == 1) { in TryMergeIntoShifterOperand() 212 if (type == DataType::Type::kInt64 in VisitArrayGet() 233 if (type == DataType::Type::kInt64 in VisitArraySet()
|
D | code_generator_vector_arm64_sve.cc | 58 case DataType::Type::kInt64: in VisitVecReplicateScalar() 111 case DataType::Type::kInt64: in VisitVecReplicateScalar() 150 case DataType::Type::kInt64: in VisitVecExtractScalar() 173 case DataType::Type::kInt64: in VisitVecExtractScalar() 204 case DataType::Type::kInt64: in CreateVecUnOpLocations() 239 case DataType::Type::kInt64: in VisitVecReduce() 297 case DataType::Type::kInt64: in VisitVecNeg() 336 case DataType::Type::kInt64: in VisitVecAbs() 373 case DataType::Type::kInt64: in VisitVecNot() 392 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | code_generator_vector_arm64_neon.cc | 58 case DataType::Type::kInt64: in VisitVecReplicateScalar() 111 case DataType::Type::kInt64: in VisitVecReplicateScalar() 150 case DataType::Type::kInt64: in VisitVecExtractScalar() 173 case DataType::Type::kInt64: in VisitVecExtractScalar() 204 case DataType::Type::kInt64: in CreateVecUnOpLocations() 239 case DataType::Type::kInt64: in VisitVecReduce() 297 case DataType::Type::kInt64: in VisitVecNeg() 336 case DataType::Type::kInt64: in VisitVecAbs() 373 case DataType::Type::kInt64: in VisitVecNot() 392 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | side_effects_test.cc | 32 DataType::Type::kInt64, 173 SideEffects::FieldWriteOfType(DataType::Type::kInt64, /* is_volatile= */ false), in TEST() 176 SideEffects::ArrayWriteOfType(DataType::Type::kInt64), in TEST() 249 SideEffects::ArrayReadOfType(DataType::Type::kInt64).ToString().c_str()); in TEST() 258 s = s.Union(SideEffects::FieldWriteOfType(DataType::Type::kInt64, /* is_volatile= */ false)); in TEST()
|
D | code_generator_vector_x86_64.cc | 39 case DataType::Type::kInt64: in VisitVecReplicateScalar() 90 case DataType::Type::kInt64: in VisitVecReplicateScalar() 120 case DataType::Type::kInt64: in VisitVecExtractScalar() 150 case DataType::Type::kInt64: in VisitVecExtractScalar() 176 case DataType::Type::kInt64: in CreateVecUnOpLocations() 191 if (instruction->GetPackedType() == DataType::Type::kInt64 || in VisitVecReduce() 218 case DataType::Type::kInt64: { in VisitVecReduce() 284 case DataType::Type::kInt64: in VisitVecNeg() 373 case DataType::Type::kInt64: in VisitVecNot() 405 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | code_generator_vector_x86.cc | 33 case DataType::Type::kInt64: in VisitVecReplicateScalar() 95 case DataType::Type::kInt64: { in VisitVecReplicateScalar() 123 case DataType::Type::kInt64: in VisitVecExtractScalar() 163 case DataType::Type::kInt64: { in VisitVecExtractScalar() 193 case DataType::Type::kInt64: in CreateVecUnOpLocations() 208 if (instruction->GetPackedType() == DataType::Type::kInt64 || in VisitVecReduce() 235 case DataType::Type::kInt64: { in VisitVecReduce() 301 case DataType::Type::kInt64: in VisitVecNeg() 390 case DataType::Type::kInt64: in VisitVecNot() 422 case DataType::Type::kInt64: in CreateVecBinOpLocations() [all …]
|
D | instruction_builder.cc | 1511 bool is_wide = (type == DataType::Type::kInt64) || (type == DataType::Type::kFloat64); in SetupInvokeArguments() 1583 constexpr DataType::Type kInt64 = DataType::Type::kInt64; in BuildSimpleIntrinsic() local 1597 instruction = new (allocator_) HRor(kInt64, /*value=*/ nullptr, /*distance=*/ nullptr); in BuildSimpleIntrinsic() 1605 kInt64, /*first=*/ nullptr, /*second=*/ nullptr, ComparisonBias::kNoBias, dex_pc); in BuildSimpleIntrinsic() 1613 kInt64, /*first=*/ nullptr, graph_->GetLongConstant(0), ComparisonBias::kNoBias, dex_pc); in BuildSimpleIntrinsic() 1669 instruction = new (allocator_) HMin(kInt64, /*left=*/ nullptr, /*right=*/ nullptr, dex_pc); in BuildSimpleIntrinsic() 1681 instruction = new (allocator_) HMax(kInt64, /*left=*/ nullptr, /*right=*/ nullptr, dex_pc); in BuildSimpleIntrinsic() 1693 instruction = new (allocator_) HAbs(kInt64, /*input=*/ nullptr, dex_pc); in BuildSimpleIntrinsic() 2041 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in BuildCheckedDivRem() 2057 || (type == DataType::Type::kInt64 && second->AsLongConstant()->GetValue() == 0)) { in BuildCheckedDivRem() [all …]
|
D | licm_test.cc | 124 DataType::Type::kInt64, in TEST_F() 151 DataType::Type::kInt64, in TEST_F() 162 DataType::Type::kInt64, in TEST_F()
|
D | data_type_test.cc | 41 CheckConversion<DataType::Type::kInt64, Primitive::kPrimLong>(); in TEST() 70 DataType::Type::kInt64, in TEST()
|
D | nodes_shared.cc | 50 if (result_type == DataType::Type::kInt32 && input_type == DataType::Type::kInt64) { in GetOpInfoFromInstruction()
|
D | code_generator_x86.cc | 1224 case DataType::Type::kInt64: in GetReturnLocation() 1260 case DataType::Type::kInt64: { in GetNextLocation() 1449 if (dst_type == DataType::Type::kInt64 && !src.IsConstant() && !src.IsFpuRegister()) { in MoveLocation() 1680 case DataType::Type::kInt64: in GenerateCompareTestAndBranch() 1710 cond->InputAt(0)->GetType() != DataType::Type::kInt64 && in AreEflagsSetFrom() 1776 if (type == DataType::Type::kInt64 || DataType::IsFloatingPointType(type)) { in GenerateTestAndBranch() 1858 if (compare_type == DataType::Type::kInt64 || in SelectCanUseCMOV() 1918 DCHECK_NE(condition->InputAt(0)->GetType(), DataType::Type::kInt64); in VisitSelect() 1933 if (select->GetType() == DataType::Type::kInt64) { in VisitSelect() 1989 case DataType::Type::kInt64: { in HandleCondition() [all …]
|
D | code_generator_x86_64.cc | 133 DCHECK_EQ(DataType::Type::kInt64, type_); in EmitNativeCode() 1072 case DataType::Type::kInt64: in GenerateStaticOrDirectCall() 1717 case DataType::Type::kInt64: { in GenerateCompareTest() 1770 case DataType::Type::kInt64: { in GenerateCompareTestAndBranch() 1866 if (type == DataType::Type::kInt64 || DataType::IsFloatingPointType(type)) { in GenerateTestAndBranch() 2060 case DataType::Type::kInt64: in HandleCondition() 2100 case DataType::Type::kInt64: in HandleCondition() 2239 case DataType::Type::kInt64: { in VisitCompare() 2277 case DataType::Type::kInt64: { in VisitCompare() 2414 case DataType::Type::kInt64: in VisitReturn() [all …]
|
D | intrinsics_arm_vixl.cc | 300 DCHECK((type == DataType::Type::kInt32) || (type == DataType::Type::kInt64)); in GenNumberOfLeadingZeros() 302 if (type == DataType::Type::kInt64) { in GenNumberOfLeadingZeros() 332 GenNumberOfLeadingZeros(invoke, DataType::Type::kInt64, codegen_); in VisitLongNumberOfLeadingZeros() 338 DCHECK((type == DataType::Type::kInt32) || (type == DataType::Type::kInt64)); in GenNumberOfTrailingZeros() 344 if (type == DataType::Type::kInt64) { in GenNumberOfTrailingZeros() 378 GenNumberOfTrailingZeros(invoke, DataType::Type::kInt64, codegen_); in VisitLongNumberOfTrailingZeros() 616 case DataType::Type::kInt64: { in GenUnsafeGet() 673 CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt64); in VisitUnsafeGetLong() 676 CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt64); in VisitUnsafeGetLongVolatile() 692 GenUnsafeGet(invoke, DataType::Type::kInt64, /* is_volatile= */ false, codegen_); in VisitUnsafeGetLong() [all …]
|
D | code_generator_arm_vixl.cc | 283 case DataType::Type::kInt64: in GetLoadOperandType() 308 case DataType::Type::kInt64: in GetStoreOperandType() 1128 DCHECK_EQ(instruction->GetType(), DataType::Type::kInt64); in GenerateLongDataProc() 1308 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateLongTestConstant() 1429 DCHECK_EQ(condition->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateLongTest() 1508 if (type == DataType::Type::kInt64) { in GenerateTest() 1554 DCHECK_EQ(cond->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateEqualLong() 1610 DCHECK_EQ(cond->GetLeft()->GetType(), DataType::Type::kInt64); in GenerateConditionLong() 1679 if (type == DataType::Type::kInt64) { in GenerateConditionIntegralOrNonPrimitive() 1770 if (type == DataType::Type::kInt64) { in CanEncodeConstantAs8BitImmediate() [all …]
|
D | intrinsics_arm64.cc | 248 case DataType::Type::kInt64: in GenReverseBytes() 270 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongReverseBytes() 284 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenNumberOfLeadingZeros() 305 GenNumberOfLeadingZeros(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongNumberOfLeadingZeros() 311 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenNumberOfTrailingZeros() 333 GenNumberOfTrailingZeros(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongNumberOfTrailingZeros() 339 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in GenReverse() 360 GenReverse(invoke->GetLocations(), DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongReverse() 372 VRegister fpr = (type == DataType::Type::kInt64) ? temps.AcquireD() : temps.AcquireS(); in GenBitCount() 385 GenBitCount(invoke, DataType::Type::kInt64, GetVIXLAssembler()); in VisitLongBitCount() [all …]
|
D | intrinsics_x86.cc | 1487 case DataType::Type::kInt64: in GenPeek() 1518 GenPeek(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPeekLongNative() 1573 case DataType::Type::kInt64: in GenPoke() 1606 CreateLongIntToVoidLocations(allocator_, DataType::Type::kInt64, invoke); in VisitMemoryPokeLongNative() 1610 GenPoke(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPokeLongNative() 1670 case DataType::Type::kInt64: { in GenUnsafeGet() 1712 if (type == DataType::Type::kInt64) { in CreateIntIntIntToIntLocations() 1735 allocator_, invoke, DataType::Type::kInt64, /* is_volatile= */ false); in VisitUnsafeGetLong() 1738 …CreateIntIntIntToIntLocations(allocator_, invoke, DataType::Type::kInt64, /* is_volatile= */ true); in VisitUnsafeGetLongVolatile() 1757 GenUnsafeGet(invoke, DataType::Type::kInt64, /* is_volatile= */ false, codegen_); in VisitUnsafeGetLong() [all …]
|
D | intrinsics_x86_64.cc | 201 case DataType::Type::kInt64: in GenReverseBytes() 223 GenReverseBytes(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitLongReverseBytes() 1650 case DataType::Type::kInt64: in GenPeek() 1680 GenPeek(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPeekLongNative() 1728 case DataType::Type::kInt64: in GenPoke() 1765 GenPoke(invoke->GetLocations(), DataType::Type::kInt64, GetAssembler()); in VisitMemoryPokeLongNative() 1824 case DataType::Type::kInt64: in GenUnsafeGet() 1881 GenUnsafeGet(invoke, DataType::Type::kInt64, /* is_volatile= */ false, codegen_); in VisitUnsafeGetLong() 1884 GenUnsafeGet(invoke, DataType::Type::kInt64, /* is_volatile= */ true, codegen_); in VisitUnsafeGetLongVolatile() 1929 CreateIntIntIntIntToVoidPlusTempsLocations(allocator_, DataType::Type::kInt64, invoke); in VisitUnsafePutLong() [all …]
|
D | constant_folding_test.cc | 202 DataType::Type::kInt64); in TEST_F() 477 DataType::Type::kInt64); in TEST_F() 541 DataType::Type::kInt64); in TEST_F()
|
D | instruction_simplifier.cc | 347 int64_t implicit_mask = (value->GetType() == DataType::Type::kInt64) in VisitShift() 851 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in NewIntegralAbs() 864 DCHECK(type == DataType::Type::kInt32 || type == DataType::Type::kInt64); in NewIntegralMinMax() 891 to_type == DataType::Type::kInt64); in AreLowerPrecisionArgs() 1094 !(result_type == DataType::Type::kInt64 && input_type == DataType::Type::kFloat32); in IsTypeConversionLossless() 1122 input_type == DataType::Type::kInt64 || in IsTypeConversionForStoringIntoNoWiderFieldOnly() 1123 result_type == DataType::Type::kInt64) { in IsTypeConversionForStoringIntoNoWiderFieldOnly() 1378 input_other->GetType() == DataType::Type::kInt64 && in VisitAnd() 1394 DCHECK_NE(new_and_input->GetType(), DataType::Type::kInt64); in VisitAnd() 1400 new (GetGraph()->GetAllocator()) HTypeConversion(DataType::Type::kInt64, new_and); in VisitAnd() [all …]
|
D | code_generator.cc | 75 return type == DataType::Type::kInt64; in CheckType() 79 return (DataType::IsIntegralType(type) && type != DataType::Type::kInt64) in CheckType() 83 return (type == DataType::Type::kInt64) || (type == DataType::Type::kFloat64); in CheckType() 86 return DataType::IsIntegralType(type) && (type != DataType::Type::kInt64); in CheckType() 90 return type == DataType::Type::kInt64; in CheckType() 770 (DataType::Is64BitType(field_type) ? DataType::Type::kInt64 in GenerateUnresolvedFieldAccess() 807 case DataType::Type::kInt64: in GenerateUnresolvedFieldAccess() 1424 if (current->GetType() == DataType::Type::kInt64) { in EmitVRegInfo() 1431 if (current->GetType() == DataType::Type::kInt64) { in EmitVRegInfo()
|
D | code_generator_arm64.cc | 139 } else if (return_type == DataType::Type::kInt64) { in ARM64ReturnLocation() 1445 (cst->IsLongConstant() && type == DataType::Type::kInt64) || in CoherentConstantAndType() 1490 dst_type = destination.IsRegister() ? DataType::Type::kInt64 : DataType::Type::kFloat64; in MoveLocation() 1510 ? DataType::Type::kInt64 in MoveLocation() 1537 dst_type = destination.IsStackSlot() ? DataType::Type::kInt32 : DataType::Type::kInt64; in MoveLocation() 1618 case DataType::Type::kInt64: in Load() 1678 case DataType::Type::kInt64: in LoadAcquire() 1727 case DataType::Type::kInt64: in Store() 1781 case DataType::Type::kInt64: in StoreRelease() 1961 case DataType::Type::kInt64: in HandleBinaryOp() [all …]
|