/art/runtime/interpreter/mterp/arm64ng/ |
D | invoke.S | 27 mov x2, xPC 81 ldr x2, [x2, #MIRROR_CLASS_IMT_PTR_OFFSET_64] 82 ldr x0, [x2, w0, uxtw #3] 91 mov x2, xPC 116 mov x2, xPC 129 FETCH_FROM_THREAD_CACHE x2, 2f 144 mov x2, xPC 146 mov x2, x0
|
D | main.S | 314 add x2, x2, #1 383 add x2, x2, #1 397 mov x2, xFP 437 SAVE_TWO_REGS x2, x3, 16 451 RESTORE_TWO_REGS x2, x3, 16 511 asr ip2, x2, #4 528 and x2, x2, #15 535 asr x2, x3, #12 542 ubfx x2, x3, #8, #4 549 ubfx x2, x3, #4, #4 [all …]
|
D | array.S | 21 ldr x2, [x0, #$data_offset] // x2<- vBB[vCC] 22 SET_VREG_WIDE x2, w9 81 GET_VREG_WIDE x2, w9 // x2<- vAA 86 $store x2, [x0, #$data_offset] // vBB[vCC]<- x2 152 mov x2, xFP
|
D | other.S | 61 mov x2, xPC 95 orr x0, x0, x2, lsl #32 // w0<- hhhhBBBBbbbb 113 FETCH_S x2, 2 // x2<- ssssssssssssBBBB (high) 116 orr x0, x0, x2, lsl #16 // x0<- ssssssssBBBBbbbb
|
D | object.S | 127 strb w1, [x1, x2]
|
/art/runtime/interpreter/mterp/arm64/ |
D | main.S | 412 str x3, [x2, #SHADOWFRAME_RESULT_REGISTER_OFFSET] 415 str x1, [x2, #SHADOWFRAME_DEX_INSTRUCTIONS_OFFSET] 419 ldr w0, [x2, #SHADOWFRAME_NUMBER_OF_VREGS_OFFSET] 420 add xFP, x2, #SHADOWFRAME_VREGS_OFFSET // point to vregs. 422 ldr w0, [x2, #SHADOWFRAME_DEX_PC_OFFSET] // Get starting dex_pc. 433 mov x2, xSELF 450 mov x2, xPC 540 ldr x2, [xSELF, #THREAD_FLAGS_OFFSET] 630 mov x2, xINST 648 mov x2, xINST [all …]
|
D | object.S | 8 add x2, xFP, #OFF_FP_SHADOWFRAME // arg2: ShadowFrame* sf 26 ldr x2, [xFP, #OFF_FP_METHOD] // w2<- method 48 ldr x0, [x2, x1] // x0<- obj.field 50 ${load} w0, [x2, x1] // w0<- obj.field 170 ldr x2, [xFP, #OFF_FP_METHOD] // w2<- method 249 str x0, [x2, x3] // obj.field<- x0
|
D | invoke.S | 11 mov x2, xPC 32 mov x2, xPC
|
D | other.S | 10 add x2, xFP, #OFF_FP_SHADOWFRAME 83 add x2, xFP, #OFF_FP_SHADOWFRAME 102 orr x0, x0, x2, lsl #32 // w0<- hhhhBBBBbbbb 120 FETCH_S x2, 2 // x2<- ssssssssssssBBBB (high) 123 orr x0, x0, x2, lsl #16 // x0<- ssssssssBBBBbbbb
|
D | array.S | 82 ldr x2, [x0, #MIRROR_WIDE_ARRAY_DATA_OFFSET] // x2<- vBB[vCC] 84 SET_VREG_WIDE x2, w4 207 mov x2, xSELF
|
D | arithmetic.S | 205 $opcode x0, x1, x2 // Do the shift. Only low 6 bits of x2 are used. 307 GET_VREG_WIDE x2, w3 308 cmp x1, x2
|
/art/runtime/arch/arm64/ |
D | quick_entrypoints_arm64.S | 116 SAVE_TWO_REGS x2, x3, 288 174 RESTORE_TWO_REGS x2, x3, 288 255 mov x2, xSELF // pass Thread::Current 338 mov x2, xSELF // pass Thread::Current 372 add x10, x2, #(__SIZEOF_POINTER__ + 0xf) // Reserve space for ArtMethod*, arguments and 392 ldr w10, [x1, x2] 393 str w10, [x9, x2] 582 LOADREG x8 8 x2 .LfillRegisters 707 LOADREG x8 8 x2 .LfillRegisters2 823 br x2 [all …]
|
D | jni_entrypoints_arm64.S | 44 stp x2, x3, [sp, #-16]! 69 ldp x2, x3, [sp], #16 103 stp x2, x3, [sp, #16] 122 ldp x2, x3, [sp, #16] 166 stp x1, x2, [x13, #80] 242 ldp x1, x2, [x29, #80]
|
D | memcmp16_arm64.S | 30 #define limit x2
|
D | asm_support_arm64.S | 246 SAVE_TWO_REGS_BASE \base, x1, x2, 80 274 RESTORE_TWO_REGS_BASE \base, x1, x2, 80
|
/art/test/661-checker-simd-reduc/src/ |
D | Main.java | 427 int[] x2 = { 1, 1, 1, 1, 0, 0, 0, 0 }; in main() local 435 expectEquals(118, reductionIntToLoop(x2)); in main()
|
/art/test/475-regression-inliner-ids/smali/ |
D | TestCase.smali | 62 0x2 -> :sswitch_8
|
/art/test/547-regression-trycatch-critic-edge/smali/ |
D | TestCase.smali | 50 const v4, 0x2
|
/art/test/593-checker-shift-and-simplifier/smali/ |
D | SmaliTests.smali | 53 shl-int/lit8 v3, v2, 0x2
|
/art/test/586-checker-null-array-get/smali/ |
D | SmaliTests.smali | 36 const/4 v5, 0x2 105 const/4 v5, 0x2
|
/art/test/121-modifiers/smali/ |
D | NonInf.smali | 59 const/4 v0, 0x2
|
/art/test/449-checker-bce/src/ |
D | Main.java | 1793 int[] x2 = new int[10]; in main() local 1796 modArrayIndex5(x2, 0); in main() 1802 if (x1[i] != e1 || x2[i] != e2 || x3[i] != e3) { in main()
|
/art/test/530-checker-loops2/src/ |
D | Main.java | 1235 byte[] x2 = { 2 }; in main() 1243 dynamicBCEAndConstantIndicesAllPrimTypes(x, x1, x2, x3, x4, x5, x6, x7, x8, 0, 10)); in main()
|
/art/test/530-checker-peel-unroll/smali/ |
D | PeelUnroll.smali | 95 rem-int/lit8 v0, v2, 0x2 204 rem-int/lit8 v0, v2, 0x2
|
/art/compiler/optimizing/ |
D | code_generator_arm64.h | 61 vixl::aarch64::x2, 160 vixl::aarch64::x2, 275 ? helpers::LocationFrom(vixl::aarch64::x2) in GetSetValueLocation()
|