Searched refs:rdx (Results 1 – 13 of 13) sorted by relevance
/art/runtime/arch/x86_64/ |
D | memcmp16_x86_64.S | 47 shl $1, %rdx 48 cmp $79, %rdx 50 add %rdx, %rsi 51 add %rdx, %rdi 52 BRANCH_TO_JMPTBL_ENTRY(L(table_64bytes), %rdx, 2) 67 add %rcx, %rdx 71 cmp $128, %rdx 74 sub $64, %rdx 95 cmp $32, %rdx 107 sub $32, %rdx [all …]
|
D | quick_entrypoints_x86_64.S | 83 PUSH rdx 164 POP rdx 234 movq %gs:THREAD_SELF_OFFSET, %rdx // pass Thread::Current() 312 movq %gs:THREAD_SELF_OFFSET, %rdx // pass Thread 318 movq %rdx, %rax 453 subq %rdx, %rsp // Reserve stack space for argument array. 471 LOOP_OVER_SHORTY_LOADING_GPRS rdx, edx, .Lgpr_setup_finished 547 subq %rdx, %rsp // Reserve stack space for argument array. 564 LOOP_OVER_SHORTY_LOADING_GPRS rdx, edx, .Lgpr_setup_finished2 638 popq %rdx [all …]
|
D | jni_entrypoints_x86_64.S | 28 PUSH_ARG rdx // Arg. 67 POP_ARG rdx // Arg. 90 PUSH_ARG rdx 124 POP_ARG rdx 176 movq %rdx, 120(%rax) 259 movq 120(%rbp), %rdx
|
D | asm_support_x86_64.S | 351 PUSH_ARG rdx // Quick arg 2. 393 POP_ARG rdx
|
/art/runtime/interpreter/mterp/x86_64ng/ |
D | other.S | 47 movq rPC, %rdx 191 SET_WIDE_VREG %rax, rINSTq # v[AA] <- rdx 200 GET_WIDE_VREG %rdx, rINSTq # rdx <- v[B] 201 SET_WIDE_VREG %rdx, %rcx # v[A] <- rdx 209 GET_WIDE_VREG %rdx, %rcx # rdx <- v[B] 210 SET_WIDE_VREG %rdx, %rax # v[A] <- rdx 217 GET_WIDE_VREG %rdx, %rcx # rdx <- v[B] 218 SET_WIDE_VREG %rdx, rINSTq # v[A] <- rdx
|
D | invoke.S | 31 movq rPC, %rdx 82 movq MIRROR_CLASS_IMT_PTR_OFFSET_64(%edx), %rdx 83 movq (%rdx, %rax, 8), %rdi 88 movq rPC, %rdx 114 movq rPC, %rdx 149 movq rPC, %rdx
|
D | main.S | 46 #define IN_ARG2 %rdx 51 #define OUT_ARG2 %rdx 760 movq rFP, %rdx 1054 LOOP_OVER_SHORTY_LOADING_GPRS rdx, edx, r11, r9, r10, .Lgpr_setup_finished_\suffix 1151 LOOP_RANGE_OVER_SHORTY_LOADING_GPRS rdx, edx, r11, r10, rbp, .Lgpr_setup_finished_range_\suffix 1197 movq rPC, %rdx 1198 salq MACRO_LITERAL(THREAD_INTERPRETER_CACHE_SIZE_SHIFT), %rdx 1199 andq MACRO_LITERAL(THREAD_INTERPRETER_CACHE_SIZE_MASK), %rdx 1200 cmpq THREAD_INTERPRETER_CACHE_OFFSET(%rax, %rdx, 1), rPC 1202 movq __SIZEOF_POINTER__+THREAD_INTERPRETER_CACHE_OFFSET(%rax, %rdx, 1), \dest_reg [all …]
|
/art/runtime/interpreter/mterp/x86_64/ |
D | other.S | 214 movq (%rax), %rdx # Get wide 215 SET_WIDE_VREG %rdx, rINSTq # v[AA] <- rdx 224 GET_WIDE_VREG %rdx, rINSTq # rdx <- v[B] 225 SET_WIDE_VREG %rdx, %rcx # v[A] <- rdx 233 GET_WIDE_VREG %rdx, %rcx # rdx <- v[B] 234 SET_WIDE_VREG %rdx, %rax # v[A] <- rdx 241 GET_WIDE_VREG %rdx, %rcx # rdx <- v[B] 242 SET_WIDE_VREG %rdx, rINSTq # v[A] <- rdx
|
D | arithmetic.S | 21 $ext # rdx:rax <- sign-extended of rax 82 $ext # rdx:rax <- sign-extended of rax 411 movzbq 2(rPC), %rdx # edx <- BB 413 GET_WIDE_VREG %rdx, %rdx # rdx <- v[BB] 418 cmpq VREG_ADDRESS(%rcx), %rdx
|
D | main.S | 146 #define IN_ARG2 %rdx 153 #define OUT_ARG2 %rdx 697 movq OFF_FP_RESULT_REGISTER(rFP), %rdx 698 movq %rax, (%rdx)
|
/art/compiler/optimizing/ |
D | code_generator_x86_64.cc | 3844 CpuRegister rdx = locations->GetTemp(0).AsRegister<CpuRegister>(); in DivByPowerOfTwo() local 3846 __ movq(rdx, numerator); in DivByPowerOfTwo() 3847 __ shrq(rdx, Immediate(63)); in DivByPowerOfTwo() 3848 __ addq(rdx, numerator); in DivByPowerOfTwo() 3850 codegen_->Load64BitValue(rdx, abs_imm - 1); in DivByPowerOfTwo() 3851 __ addq(rdx, numerator); in DivByPowerOfTwo() 3853 __ cmov(kGreaterEqual, rdx, numerator); in DivByPowerOfTwo() 3856 __ sarq(rdx, Immediate(shift)); in DivByPowerOfTwo() 3859 __ negq(rdx); in DivByPowerOfTwo() 3862 __ movq(output_register, rdx); in DivByPowerOfTwo() [all …]
|
D | intrinsics_x86_64.cc | 2703 CpuRegister rdx = locations->GetTemp(0).AsRegister<CpuRegister>(); in VisitIntegerDivideUnsigned() local 2708 DCHECK_EQ(RDX, rdx.AsRegister()); in VisitIntegerDivideUnsigned() 2716 __ xorl(rdx, rdx); in VisitIntegerDivideUnsigned()
|
/art/runtime/arch/ |
D | stub_test.cc | 335 PUSH(%%rdx) in Invoke3WithReferrerAndHidden() 359 POP(%%rdx) in Invoke3WithReferrerAndHidden() 378 POP(%%rdx) in Invoke3WithReferrerAndHidden()
|