/external/compiler-rt/lib/i386/ |
H A D | ashrdi3.S | 18 movd 8(%esp), %xmm1 19 punpckldq %xmm1, %xmm0 // Load input 31 pcmpeqb %xmm1, %xmm1 // -1ULL 32 psrlq $58, %xmm1 // 0x3f 33 pandn %xmm1, %xmm2 // 63 - count 34 pcmpeqb %xmm1, %xmm1 // -1ULL 35 psubq %xmm1, %xmm2 // 64 - count 36 psllq %xmm2, %xmm1 // [all...] |
H A D | floatdidf.S | 22 cvtsi2sd 8(%esp), %xmm1 26 mulsd REL_ADDR(twop32), %xmm1 // a_hi as a double (without rounding) 28 subsd %xmm2, %xmm1 // a_hi - 0x1p52 (no rounding occurs) 30 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
H A D | floatdisf.S | 22 movd 8(%esp), %xmm1 23 punpckldq %xmm1, %xmm0
|
H A D | floatdixf.S | 22 movd 8(%esp), %xmm1 23 punpckldq %xmm1, %xmm0
|
H A D | floatundidf.S | 34 movss 8(%esp), %xmm1 // high 32 bits of a 38 orpd REL_ADDR(twop84), %xmm1 // 0x1p84 + a_hi (no rounding occurs) 39 subsd REL_ADDR(twop84_plus_twop52), %xmm1 // a_hi - 0x1p52 (no rounding occurs) 41 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
H A D | floatundixf.S | 27 movss 4(%esp), %xmm1 // lo 32 bits of input 29 orpd REL_ADDR(twop52), %xmm1 // 2^52 + lo (as a double) 31 movsd %xmm1, 4(%esp)
|
H A D | floatundisf.S | 33 movd 8(%esp), %xmm1 35 punpckldq %xmm1, %xmm0 73 movd 8(%esp), %xmm1 75 punpckldq %xmm1, %xmm0 82 movsd STICKY, %xmm1 // (big input) ? 0xfff : 0 86 andpd %xmm0, %xmm1 // (big input) ? input & 0xfff : 0 89 orpd %xmm2, %xmm1 // 0x1.0p52 + ((big input) ? input & 0xfff : input) 90 orpd %xmm1, %xmm0 // 0x1.0p52 + ((big input) ? (input >> 12 | input & 0xfff) : input)
|
/external/llvm/test/MC/X86/ |
H A D | x86_64-fma4-encoding.s | 4 // CHECK: vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 8 // CHECK: vfmaddss %xmm1, (%rcx), %xmm0, %xmm0 10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0 12 // CHECK: vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 16 // CHECK: vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0 18 vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0 20 // CHECK: vfmaddsd %xmm1, (%rcx), %xmm0, %xmm0 22 vfmaddsd %xmm1, ( [all...] |
H A D | x86_64-xop-encoding.s | 8 // CHECK: vphsubwd (%rcx,%rax), %xmm1 10 vphsubwd (%rcx,%rax), %xmm1 11 // CHECK: vphsubwd %xmm0, %xmm1 13 vphsubwd %xmm0, %xmm1 16 // CHECK: vphsubdq (%rcx,%rax), %xmm1 18 vphsubdq (%rcx,%rax), %xmm1 19 // CHECK: vphsubdq %xmm0, %xmm1 21 vphsubdq %xmm0, %xmm1 24 // CHECK: vphsubbw (%rax), %xmm1 26 vphsubbw (%rax), %xmm1 [all...] |
H A D | x86-32-fma3.s | 3 // CHECK: vfmadd132pd %xmm2, %xmm5, %xmm1 5 vfmadd132pd %xmm2, %xmm5, %xmm1 7 // CHECK: vfmadd132pd (%eax), %xmm5, %xmm1 9 vfmadd132pd (%eax), %xmm5, %xmm1 11 // CHECK: vfmadd132ps %xmm2, %xmm5, %xmm1 13 vfmadd132ps %xmm2, %xmm5, %xmm1 15 // CHECK: vfmadd132ps (%eax), %xmm5, %xmm1 17 vfmadd132ps (%eax), %xmm5, %xmm1 19 // CHECK: vfmadd213pd %xmm2, %xmm5, %xmm1 21 vfmadd213pd %xmm2, %xmm5, %xmm1 [all...] |
H A D | x86-32-avx.s | 274 // CHECK: vunpckhps %xmm1, %xmm2, %xmm4 276 vunpckhps %xmm1, %xmm2, %xmm4 278 // CHECK: vunpckhpd %xmm1, %xmm2, %xmm4 280 vunpckhpd %xmm1, %xmm2, %xmm4 282 // CHECK: vunpcklps %xmm1, %xmm2, %xmm4 284 vunpcklps %xmm1, %xmm2, %xmm4 286 // CHECK: vunpcklpd %xmm1, %xmm2, %xmm4 288 vunpcklpd %xmm1, %xmm2, %xmm4 306 // CHECK: vcmpps $0, %xmm0, %xmm6, %xmm1 308 vcmpps $0, %xmm0, %xmm6, %xmm1 [all...] |
/external/libvpx/vp8/encoder/x86/ |
H A D | variance_impl_ssse3.asm | 68 movdqu xmm1, XMMWORD PTR [rsi+1] 71 punpcklbw xmm0, xmm1 72 punpckhbw xmm2, xmm1 92 movdqu xmm1, XMMWORD PTR [rsi] 94 movdqa xmm3, xmm1 96 punpcklbw xmm1, xmm2 98 pmaddubsw xmm1, [rax] 101 paddw xmm1, [GLOBAL(xmm_bi_rd)] 103 psraw xmm1, xmm_filter_shift 105 packuswb xmm1, xmm [all...] |
H A D | dct_sse2.asm | 70 movq xmm1, MMWORD PTR[input ] ;23 22 21 20 74 punpcklqdq xmm1, xmm3 ;33 32 31 30 23 22 21 20 77 punpckldq xmm0, xmm1 ;23 22 03 02 21 20 01 00 78 punpckhdq xmm2, xmm1 ;33 32 13 12 31 30 11 10 79 movdqa xmm1, xmm0 81 pshufhw xmm1, xmm1, 0b1h ;22 23 02 03 xx xx xx xx 84 punpckhdq xmm1, xmm2 ;32 33 22 23 12 13 02 03 86 paddw xmm0, xmm1 ;b1 a1 b1 a1 b1 a1 b1 a1 87 psubw xmm3, xmm1 ;c [all...] |
H A D | variance_impl_sse2.asm | 38 movdqa xmm1, [rax+16] 42 pmaddwd xmm1, xmm1 46 paddd xmm0, xmm1 129 movdqu xmm1, XMMWORD PTR [rsi] 135 movdqa xmm3, xmm1 139 punpcklbw xmm1, xmm0 146 psubw xmm1, xmm2 149 paddw xmm7, xmm1 150 pmaddwd xmm1, xmm [all...] |
H A D | subtract_sse2.asm | 96 movdqa xmm1, XMMWORD PTR [rax] ; pred 99 psubb xmm0, xmm1 101 pxor xmm1, [GLOBAL(t80)] ;convert to signed values 103 pcmpgtb xmm1, xmm2 ; obtain sign information 106 movdqa xmm3, xmm1 107 punpcklbw xmm0, xmm1 ; put sign back to subtraction 171 movdqa xmm1, XMMWORD PTR [rax] ; pred 175 psubb xmm0, xmm1 ; subtraction with sign missed 177 pxor xmm1, [GLOBAL(t80)] ;convert to signed values 179 pcmpgtb xmm1, xmm [all...] |
H A D | quantize_ssse3.asm | 39 movdqa xmm1, xmm0 45 pabsw xmm1, xmm1 48 paddw xmm1, xmm2 51 pmulhw xmm1, [rsi] 58 pxor xmm1, xmm0 60 psubw xmm1, xmm0 63 movdqa [rdi], xmm1 70 pmullw xmm2, xmm1 73 pcmpeqw xmm1, xmm [all...] |
H A D | fwalsh_sse2.asm | 32 movq xmm1, MMWORD PTR [rsi + rdx] 37 punpcklwd xmm0, xmm1 40 movdqa xmm1, xmm0 42 punpckhdq xmm1, xmm2 ; ip[3] ip[2] 45 paddw xmm0, xmm1 46 psubw xmm2, xmm1 51 movdqa xmm1, xmm0 53 punpckhqdq xmm1, xmm2 ; c1 d1 62 paddw xmm0, xmm1 ; b1+c1 a1+d1 63 psubw xmm2, xmm1 ; b [all...] |
H A D | quantize_sse2.asm | 69 movdqa xmm1, xmm0 77 pxor xmm1, xmm0 81 psubw xmm1, xmm0 93 psubw xmm1, xmm2 95 movdqa [rsp + abs_minus_zbin], xmm1 99 paddw xmm1, xmm2 109 paddw xmm1, xmm2 113 pmulhw xmm3, xmm1 117 paddw xmm1, xmm3 120 movdqa [rsp + temp_qcoeff], xmm1 [all...] |
H A D | temporal_filter_apply_sse2.asm | 81 movq xmm1, [rsi] ; second row 83 punpcklbw xmm1, xmm7 ; src[ 8-15] 89 movdqa xmm1, xmm0 91 punpckhbw xmm1, xmm7 ; src[ 8-15] 101 psubw xmm1, xmm3 ; src - pred[ 8-15] 105 pmullw xmm1, xmm1 ; modifer[ 8-15]^2 109 pmullw xmm1, [GLOBAL(_const_3w)] 113 paddw xmm1, [rsp + rounding_bit] 117 psrlw xmm1, [rs [all...] |
/external/compiler-rt/lib/x86_64/ |
H A D | floatundidf.S | 38 movd %rdi, %xmm1 39 subsd REL_ADDR(twop84_plus_twop52), %xmm1 // a_hi - 0x1p52 (no rounding occurs) 40 addsd %xmm1, %xmm0 // a_hi + a_lo (round happens here)
|
/external/libvpx/vp8/common/x86/ |
H A D | idctllm_sse2.asm | 60 movq xmm1, [rax+rcx] 66 punpcklbw xmm1, xmm7 75 paddw xmm1, xmm4 81 packuswb xmm1, xmm7 87 movq [rax + rdx], xmm1 122 ; note the transpose of xmm1 and xmm2, necessary for shuffle 126 movdqa xmm1, [rax+32] 138 pmullw xmm1, [rdx] 143 punpckldq xmm0, xmm1 144 punpckhdq xmm4, xmm1 [all...] |
H A D | postproc_sse2.asm | 66 movdqa xmm1, xmm3 ; mm1 = p0..p3 74 movdqa xmm7, xmm1 ; mm7 = r0 p0..p3 76 psubusw xmm5, xmm1 ; mm5 = r1 p0..p3 - r0 p0..p3 85 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 87 psubusw xmm5, xmm1 ; mm5 = r2 p0..p3 - r2 p0..p3 99 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 101 psubusw xmm5, xmm1 ; mm5 = r-2 p0..p3 - p0..p3 111 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 113 psubusw xmm4, xmm1 ; mm5 = r-1 p0..p3 - p0..p3 122 pand xmm1, xmm [all...] |
H A D | iwalsh_sse2.asm | 30 movdqa xmm1, [rsi + 16] ;ip[12] ip[8] 35 pshufd xmm2, xmm1, 4eh ;ip[8] ip[12] 46 movdqa xmm1, xmm4 ;c1 b1 48 psubw xmm0, xmm1 ;d1-c1 a1-b1 aka op[12] op[8] 62 movdqa xmm1, xmm4 ; 23 03 22 02 21 01 20 00 64 punpckhwd xmm1, xmm3 ; 33 23 13 03 32 22 12 02 66 pshufd xmm2, xmm1, 4eh ;ip[8] ip[12] 78 movdqa xmm1, xmm5 ;c1 b1 80 psubw xmm4, xmm1 ;d1-c1 a1-b1 aka op[12] op[8] 89 movdqa xmm1, xmm [all...] |
/external/llvm/test/MC/ELF/ |
H A D | undef.s | 20 movsd .Lsym8(%rip), %xmm1
|
H A D | merge.s | 10 movsd .Lfoo(%rip), %xmm1
|