Lines Matching refs:__

26 #define __ down_cast<X86Assembler*>(GetAssembler())->  // NOLINT  macro
69 cpu_has_avx ? __ vxorps(dst, dst, dst) : __ xorps(dst, dst); in VisitVecReplicateScalar()
78 __ movd(dst, locations->InAt(0).AsRegister<Register>()); in VisitVecReplicateScalar()
79 __ punpcklbw(dst, dst); in VisitVecReplicateScalar()
80 __ punpcklwd(dst, dst); in VisitVecReplicateScalar()
81 __ pshufd(dst, dst, Immediate(0)); in VisitVecReplicateScalar()
86 __ movd(dst, locations->InAt(0).AsRegister<Register>()); in VisitVecReplicateScalar()
87 __ punpcklwd(dst, dst); in VisitVecReplicateScalar()
88 __ pshufd(dst, dst, Immediate(0)); in VisitVecReplicateScalar()
92 __ movd(dst, locations->InAt(0).AsRegister<Register>()); in VisitVecReplicateScalar()
93 __ pshufd(dst, dst, Immediate(0)); in VisitVecReplicateScalar()
98 __ movd(dst, locations->InAt(0).AsRegisterPairLow<Register>()); in VisitVecReplicateScalar()
99 __ movd(tmp, locations->InAt(0).AsRegisterPairHigh<Register>()); in VisitVecReplicateScalar()
100 __ punpckldq(dst, tmp); in VisitVecReplicateScalar()
101 __ punpcklqdq(dst, dst); in VisitVecReplicateScalar()
107 __ shufps(dst, dst, Immediate(0)); in VisitVecReplicateScalar()
112 __ shufpd(dst, dst, Immediate(0)); in VisitVecReplicateScalar()
161 __ movd(locations->Out().AsRegister<Register>(), src); in VisitVecExtractScalar()
166 __ movd(locations->Out().AsRegisterPairLow<Register>(), src); in VisitVecExtractScalar()
167 __ pshufd(tmp, src, Immediate(1)); in VisitVecExtractScalar()
168 __ movd(locations->Out().AsRegisterPairHigh<Register>(), tmp); in VisitVecExtractScalar()
224 __ movaps(dst, src); in VisitVecReduce()
225 __ phaddd(dst, dst); in VisitVecReduce()
226 __ phaddd(dst, dst); in VisitVecReduce()
240 __ movaps(tmp, src); in VisitVecReduce()
241 __ movaps(dst, src); in VisitVecReduce()
242 __ punpckhqdq(tmp, tmp); in VisitVecReduce()
243 __ paddq(dst, tmp); in VisitVecReduce()
269 __ cvtdq2ps(dst, src); in VisitVecCnv()
287 __ pxor(dst, dst); in VisitVecNeg()
288 __ psubb(dst, src); in VisitVecNeg()
293 __ pxor(dst, dst); in VisitVecNeg()
294 __ psubw(dst, src); in VisitVecNeg()
298 __ pxor(dst, dst); in VisitVecNeg()
299 __ psubd(dst, src); in VisitVecNeg()
303 __ pxor(dst, dst); in VisitVecNeg()
304 __ psubq(dst, src); in VisitVecNeg()
308 __ xorps(dst, dst); in VisitVecNeg()
309 __ subps(dst, src); in VisitVecNeg()
313 __ xorpd(dst, dst); in VisitVecNeg()
314 __ subpd(dst, src); in VisitVecNeg()
338 __ movaps(dst, src); in VisitVecAbs()
339 __ pxor(tmp, tmp); in VisitVecAbs()
340 __ pcmpgtd(tmp, dst); in VisitVecAbs()
341 __ pxor(dst, tmp); in VisitVecAbs()
342 __ psubd(dst, tmp); in VisitVecAbs()
347 __ pcmpeqb(dst, dst); // all ones in VisitVecAbs()
348 __ psrld(dst, Immediate(1)); in VisitVecAbs()
349 __ andps(dst, src); in VisitVecAbs()
353 __ pcmpeqb(dst, dst); // all ones in VisitVecAbs()
354 __ psrlq(dst, Immediate(1)); in VisitVecAbs()
355 __ andpd(dst, src); in VisitVecAbs()
379 __ pxor(dst, dst); in VisitVecNot()
380 __ pcmpeqb(tmp, tmp); // all ones in VisitVecNot()
381 __ psubb(dst, tmp); // 16 x one in VisitVecNot()
382 __ pxor(dst, src); in VisitVecNot()
393 __ pcmpeqb(dst, dst); // all ones in VisitVecNot()
394 __ pxor(dst, src); in VisitVecNot()
398 __ pcmpeqb(dst, dst); // all ones in VisitVecNot()
399 __ xorps(dst, src); in VisitVecNot()
403 __ pcmpeqb(dst, dst); // all ones in VisitVecNot()
404 __ xorpd(dst, src); in VisitVecNot()
476 cpu_has_avx ? __ vpaddb(dst, other_src, src) : __ paddb(dst, src); in VisitVecAdd()
481 cpu_has_avx ? __ vpaddw(dst, other_src, src) : __ paddw(dst, src); in VisitVecAdd()
485 cpu_has_avx ? __ vpaddd(dst, other_src, src) : __ paddd(dst, src); in VisitVecAdd()
489 cpu_has_avx ? __ vpaddq(dst, other_src, src) : __ paddq(dst, src); in VisitVecAdd()
493 cpu_has_avx ? __ vaddps(dst, other_src, src) : __ addps(dst, src); in VisitVecAdd()
497 cpu_has_avx ? __ vaddpd(dst, other_src, src) : __ addpd(dst, src); in VisitVecAdd()
517 __ paddusb(dst, src); in VisitVecSaturationAdd()
521 __ paddsb(dst, src); in VisitVecSaturationAdd()
525 __ paddusw(dst, src); in VisitVecSaturationAdd()
529 __ paddsw(dst, src); in VisitVecSaturationAdd()
552 __ pavgb(dst, src); in VisitVecHalvingAdd()
556 __ pavgw(dst, src); in VisitVecHalvingAdd()
583 cpu_has_avx ? __ vpsubb(dst, other_src, src) : __ psubb(dst, src); in VisitVecSub()
588 cpu_has_avx ? __ vpsubw(dst, other_src, src) : __ psubw(dst, src); in VisitVecSub()
592 cpu_has_avx ? __ vpsubd(dst, other_src, src) : __ psubd(dst, src); in VisitVecSub()
596 cpu_has_avx ? __ vpsubq(dst, other_src, src) : __ psubq(dst, src); in VisitVecSub()
600 cpu_has_avx ? __ vsubps(dst, other_src, src) : __ subps(dst, src); in VisitVecSub()
604 cpu_has_avx ? __ vsubpd(dst, other_src, src) : __ subpd(dst, src); in VisitVecSub()
624 __ psubusb(dst, src); in VisitVecSaturationSub()
628 __ psubsb(dst, src); in VisitVecSaturationSub()
632 __ psubusw(dst, src); in VisitVecSaturationSub()
636 __ psubsw(dst, src); in VisitVecSaturationSub()
663 cpu_has_avx ? __ vpmullw(dst, other_src, src) : __ pmullw(dst, src); in VisitVecMul()
667 cpu_has_avx ? __ vpmulld(dst, other_src, src) : __ pmulld(dst, src); in VisitVecMul()
671 cpu_has_avx ? __ vmulps(dst, other_src, src) : __ mulps(dst, src); in VisitVecMul()
675 cpu_has_avx ? __ vmulpd(dst, other_src, src) : __ mulpd(dst, src); in VisitVecMul()
701 cpu_has_avx ? __ vdivps(dst, other_src, src) : __ divps(dst, src); in VisitVecDiv()
705 cpu_has_avx ? __ vdivpd(dst, other_src, src) : __ divpd(dst, src); in VisitVecDiv()
725 __ pminub(dst, src); in VisitVecMin()
729 __ pminsb(dst, src); in VisitVecMin()
733 __ pminuw(dst, src); in VisitVecMin()
737 __ pminsw(dst, src); in VisitVecMin()
741 __ pminud(dst, src); in VisitVecMin()
745 __ pminsd(dst, src); in VisitVecMin()
750 __ minps(dst, src); in VisitVecMin()
754 __ minpd(dst, src); in VisitVecMin()
774 __ pmaxub(dst, src); in VisitVecMax()
778 __ pmaxsb(dst, src); in VisitVecMax()
782 __ pmaxuw(dst, src); in VisitVecMax()
786 __ pmaxsw(dst, src); in VisitVecMax()
790 __ pmaxud(dst, src); in VisitVecMax()
794 __ pmaxsd(dst, src); in VisitVecMax()
799 __ maxps(dst, src); in VisitVecMax()
803 __ maxpd(dst, src); in VisitVecMax()
836 cpu_has_avx ? __ vpand(dst, other_src, src) : __ pand(dst, src); in VisitVecAnd()
840 cpu_has_avx ? __ vandps(dst, other_src, src) : __ andps(dst, src); in VisitVecAnd()
844 cpu_has_avx ? __ vandpd(dst, other_src, src) : __ andpd(dst, src); in VisitVecAnd()
877 cpu_has_avx ? __ vpandn(dst, other_src, src) : __ pandn(dst, src); in VisitVecAndNot()
881 cpu_has_avx ? __ vandnps(dst, other_src, src) : __ andnps(dst, src); in VisitVecAndNot()
885 cpu_has_avx ? __ vandnpd(dst, other_src, src) : __ andnpd(dst, src); in VisitVecAndNot()
918 cpu_has_avx ? __ vpor(dst, other_src, src) : __ por(dst, src); in VisitVecOr()
922 cpu_has_avx ? __ vorps(dst, other_src, src) : __ orps(dst, src); in VisitVecOr()
926 cpu_has_avx ? __ vorpd(dst, other_src, src) : __ orpd(dst, src); in VisitVecOr()
959 cpu_has_avx ? __ vpxor(dst, other_src, src) : __ pxor(dst, src); in VisitVecXor()
963 cpu_has_avx ? __ vxorps(dst, other_src, src) : __ xorps(dst, src); in VisitVecXor()
967 cpu_has_avx ? __ vxorpd(dst, other_src, src) : __ xorpd(dst, src); in VisitVecXor()
1006 __ psllw(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecShl()
1010 __ pslld(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecShl()
1014 __ psllq(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecShl()
1035 __ psraw(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecShr()
1039 __ psrad(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecShr()
1060 __ psrlw(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecUShr()
1064 __ psrld(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecUShr()
1068 __ psrlq(dst, Immediate(static_cast<uint8_t>(value))); in VisitVecUShr()
1121 cpu_has_avx ? __ vxorps(dst, dst, dst) : __ xorps(dst, dst); in VisitVecSetScalars()
1139 __ movd(dst, locations->InAt(0).AsRegister<Register>()); in VisitVecSetScalars()
1144 __ xorps(tmp, tmp); in VisitVecSetScalars()
1145 __ movd(dst, locations->InAt(0).AsRegisterPairLow<Register>()); in VisitVecSetScalars()
1146 __ movd(tmp, locations->InAt(0).AsRegisterPairHigh<Register>()); in VisitVecSetScalars()
1147 __ punpckldq(dst, tmp); in VisitVecSetScalars()
1152 __ movss(dst, locations->InAt(1).AsFpuRegister<XmmRegister>()); in VisitVecSetScalars()
1156 __ movsd(dst, locations->InAt(1).AsFpuRegister<XmmRegister>()); in VisitVecSetScalars()
1223 __ movaps(tmp, right); in VisitVecDotProd()
1224 __ pmaddwd(tmp, left); in VisitVecDotProd()
1225 __ paddd(acc, tmp); in VisitVecDotProd()
1227 __ vpmaddwd(tmp, left, right); in VisitVecDotProd()
1228 __ vpaddd(acc, acc, tmp); in VisitVecDotProd()
1311 __ testb(Address(locations->InAt(0).AsRegister<Register>(), count_offset), Immediate(1)); in VisitVecLoad()
1312 __ j(kNotZero, &not_compressed); in VisitVecLoad()
1314 __ movsd(reg, VecAddress(locations, 1, instruction->IsStringCharAt())); in VisitVecLoad()
1315 __ pxor(tmp, tmp); in VisitVecLoad()
1316 __ punpcklbw(reg, tmp); in VisitVecLoad()
1317 __ jmp(&done); in VisitVecLoad()
1319 __ Bind(&not_compressed); in VisitVecLoad()
1320 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad()
1321 __ Bind(&done); in VisitVecLoad()
1332 is_aligned16 ? __ movdqa(reg, address) : __ movdqu(reg, address); in VisitVecLoad()
1336 is_aligned16 ? __ movaps(reg, address) : __ movups(reg, address); in VisitVecLoad()
1340 is_aligned16 ? __ movapd(reg, address) : __ movupd(reg, address); in VisitVecLoad()
1368 is_aligned16 ? __ movdqa(address, reg) : __ movdqu(address, reg); in VisitVecStore()
1372 is_aligned16 ? __ movaps(address, reg) : __ movups(address, reg); in VisitVecStore()
1376 is_aligned16 ? __ movapd(address, reg) : __ movupd(address, reg); in VisitVecStore()
1414 #undef __