/external/llvm/test/MC/X86/ |
H A D | x86_64-xop-encoding.s | 35 // CHECK: vphaddwq %xmm6, %xmm2 37 vphaddwq %xmm6, %xmm2 48 // CHECK: vphadduwq (%rcx,%rax), %xmm6 50 vphadduwq (%rcx,%rax), %xmm6 67 // CHECK: vphaddudq %xmm6, %xmm2 69 vphaddudq %xmm6, %xmm2 107 // CHECK: vphaddbw %xmm5, %xmm6 109 vphaddbw %xmm5, %xmm6 147 // CHECK: vfrczps %xmm6, %xmm5 149 vfrczps %xmm6, [all...] |
H A D | x86-32-avx.s | 3 // CHECK: vaddss %xmm4, %xmm6, %xmm2 5 vaddss %xmm4, %xmm6, %xmm2 7 // CHECK: vmulss %xmm4, %xmm6, %xmm2 9 vmulss %xmm4, %xmm6, %xmm2 11 // CHECK: vsubss %xmm4, %xmm6, %xmm2 13 vsubss %xmm4, %xmm6, %xmm2 15 // CHECK: vdivss %xmm4, %xmm6, %xmm2 17 vdivss %xmm4, %xmm6, %xmm2 19 // CHECK: vaddsd %xmm4, %xmm6, %xmm2 21 vaddsd %xmm4, %xmm6, [all...] |
H A D | x86_64-avx-encoding.s | 403 // CHECK: vcmpps $7, -4(%rbx,%rcx,8), %xmm6, %xmm2 405 vcmpordps -4(%rbx,%rcx,8), %xmm6, %xmm2 467 // CHECK: vcmppd $7, -4(%rbx,%rcx,8), %xmm6, %xmm2 469 vcmpordpd -4(%rbx,%rcx,8), %xmm6, %xmm2 531 // CHECK: vcmpss $7, -4(%rbx,%rcx,8), %xmm6, %xmm2 533 vcmpordss -4(%rbx,%rcx,8), %xmm6, %xmm2 595 // CHECK: vcmpsd $7, -4(%rbx,%rcx,8), %xmm6, %xmm2 597 vcmpordsd -4(%rbx,%rcx,8), %xmm6, %xmm2 723 // CHECK: vcmpps $14, -4(%rbx,%rcx,8), %xmm6, %xmm2 725 vcmpgtps -4(%rbx,%rcx,8), %xmm6, [all...] |
/external/libvpx/vp8/encoder/x86/ |
H A D | variance_impl_sse2.asm | 125 pxor xmm6, xmm6 ; clear xmm6 for accumulating sse 155 paddd xmm6, xmm1 156 paddd xmm6, xmm3 165 movdqa xmm1, xmm6 166 pxor xmm6, xmm6 169 punpcklwd xmm6, xmm7 174 psrad xmm6, 1 [all...] |
H A D | fwalsh_sse2.asm | 55 pxor xmm6, xmm6 56 movq xmm6, xmm0 58 pcmpeqw xmm7, xmm6 86 pshufd xmm6, xmm1, 0x72 ; d13 d12 a13 a12 92 movdqa xmm1, xmm6 94 punpckhqdq xmm6, xmm7 ; c13 c12 d13 d12 100 paddd xmm1, xmm6 ; b23 b22 a23 a22 101 psubd xmm3, xmm6 ; c23 c22 d23 d22 110 pxor xmm6, xmm [all...] |
H A D | variance_impl_ssse3.asm | 43 pxor xmm6, xmm6 129 paddw xmm6, xmm2 130 paddw xmm6, xmm3 195 paddw xmm6, xmm1 196 paddw xmm6, xmm2 237 paddw xmm6, xmm1 238 paddw xmm6, xmm2 286 paddw xmm6, xmm1 287 paddw xmm6, xmm [all...] |
H A D | sad_ssse3.asm | 18 lddqu xmm6, XMMWORD PTR [rdi+1] 22 psadbw xmm6, xmm0 35 paddw xmm6, xmm2 51 paddw xmm6, xmm2 64 movdqa xmm6, xmm7 65 palignr xmm6, xmm4, (%2+1) 70 psadbw xmm6, xmm0 90 paddw xmm6, xmm2 113 paddw xmm6, xmm2 240 movq xmm0, xmm6 [all...] |
H A D | quantize_sse2.asm | 103 movdqa xmm6, [rdx + 16] 110 paddw xmm5, xmm6 123 pxor xmm6, xmm6 125 movdqa [rsp + qcoeff], xmm6 126 movdqa [rsp + qcoeff + 16], xmm6 204 pcmpeqw xmm2, xmm6 205 pcmpeqw xmm3, xmm6 207 pcmpeqw xmm6, xmm6 [all...] |
H A D | sad_sse3.asm | 163 lddqu xmm6, XMMWORD PTR [%3+1] 167 psadbw xmm6, xmm0 180 paddw xmm6, xmm2 198 paddw xmm6, xmm2 258 lddqu xmm6, XMMWORD PTR [%5] 263 psadbw xmm6, xmm0 278 paddw xmm6, xmm3 295 paddw xmm6, xmm3 397 movq xmm0, xmm6 398 psrldq xmm6, [all...] |
H A D | temporal_filter_apply_sse2.asm | 46 movd xmm6, arg(4) 47 movdqa [rsp + strength], xmm6 ; where strength is used, all 16 bytes are read 164 movdqa xmm6, [rdi+32] 169 paddw xmm6, xmm1 174 movdqa [rdi+32], xmm6
|
H A D | ssim_opt.asm | 90 movdqu xmm6, [rdi] 92 movdqa xmm4, xmm6 99 movdqa xmm4, xmm6 178 movq xmm6, [rdi] 181 movdqa xmm4, xmm6
|
H A D | subtract_sse2.asm | 116 movdqa xmm6, xmm4 120 pxor xmm6, [GLOBAL(t80)] 121 pcmpgtb xmm5, xmm6 ; obtain sign information 123 movdqa xmm6, xmm4 126 punpckhbw xmm6, xmm7 ; put sign back to subtraction 129 movdqa XMMWORD PTR [rdi +48], xmm6
|
/external/libvpx/vp8/common/x86/ |
H A D | idctllm_sse2.asm | 182 movdqa xmm6, xmm2 ; a1 190 psubw xmm6, xmm3 ;3 198 punpcklwd xmm4, xmm6 ; 015 011 014 010 013 009 012 008 199 punpckhwd xmm5, xmm6 ; 115 111 114 110 113 109 112 108 206 movdqa xmm6, xmm7 ; 107 103 106 102 105 101 104 100 208 punpckhdq xmm6, xmm5 ; 115 111 107 103 114 110 106 102 216 punpckldq xmm1, xmm6 ; 114 110 014 010 106 102 006 002 217 punpckhdq xmm7, xmm6 ; 115 111 015 011 107 103 007 003 254 movdqa xmm6, xmm2 ; a1 262 psubw xmm6, xmm [all...] |
H A D | postproc_sse2.asm | 85 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 86 psubusw xmm6, xmm5 ; mm6 = r0 p0..p3 - r2 p0..p3 88 paddusw xmm6, xmm5 ; mm6 = abs(r0 p0..p3 - r2 p0..p3) 89 pcmpgtw xmm6, xmm2 90 por xmm7, xmm6 ; accumulate thresholds 99 movdqa xmm6, xmm1 ; mm6 = r0 p0..p3 100 psubusw xmm6, xmm5 ; mm6 = p0..p3 - r-2 p0..p3 102 paddusw xmm6, xmm5 ; mm6 = abs(r0 p0..p3 - r-2 p0..p3) 103 pcmpgtw xmm6, xmm2 104 por xmm7, xmm6 ; accumulat [all...] |
H A D | subpixel_ssse3.asm | 62 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 89 pmaddubsw xmm2, xmm6 116 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 144 pmaddubsw xmm2, xmm6 205 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 230 pmaddubsw xmm2, xmm6 248 pmaddubsw xmm2, xmm6 280 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 297 pmaddubsw xmm2, xmm6 306 pmaddubsw xmm3, xmm6 [all...] |
H A D | loopfilter_sse2.asm | 42 movdqa xmm6, xmm1 ; q2 46 psubusb xmm2, xmm6 ; q3-=q2 48 psubusb xmm4, xmm6 ; q1-=q2 49 psubusb xmm6, xmm3 ; q2-=q1 51 por xmm4, xmm6 ; abs(q2-q1) 68 movdqa xmm6, [rsi+2*rax] ; p1 72 movlps xmm6, [rsi + rcx] ; p1 76 movhps xmm6, [rdi + rcx] 79 movdqa XMMWORD PTR [rsp + 48], xmm6 ; store p1 83 movdqa xmm3, xmm6 ; p [all...] |
H A D | subpixel_sse2.asm | 70 movdqa xmm6, xmm1 84 psrldq xmm6, 3 ; xx xx xx 0d 0c 0b 0a 09 08 07 06 05 04 03 02 01 88 punpcklbw xmm6, xmm0 ; xx08 xx07 xx06 xx05 xx04 xx03 xx02 xx01 91 pmullw xmm6, [rdx+48] ; x[ 1] * h[ 1] ; Tap 4 107 paddsw xmm4, xmm6 195 movdqa xmm6, xmm1 209 psrldq xmm6, 3 ; xx xx xx 0d 0c 0b 0a 09 08 07 06 05 04 03 02 01 213 punpcklbw xmm6, xmm0 ; xx08 xx07 xx06 xx05 xx04 xx03 xx02 xx01 216 pmullw xmm6, [rdx+48] ; x[ 1] * h[ 1] ; Tap 4 231 paddsw xmm4, xmm6 [all...] |
H A D | postproc_mmx.c | 477 movdqa xmm6, xmm1 ; local 479 psubusw xmm6, xmm5 ; 483 paddusw xmm6, xmm5 ; 485 pcmpgtw xmm6, xmm2 486 por xmm7, xmm6 ; local 500 movdqa xmm6, xmm1 ; local 502 psubusw xmm6, xmm5 ; 506 paddusw xmm6, xmm5 ; 508 pcmpgtw xmm6, xmm2 509 por xmm7, xmm6 ; local 521 movdqa xmm6, xmm1 ; local 530 por xmm7, xmm6 ; local 611 movdqa xmm6, xmm1 ; local 620 por xmm7, xmm6 ; local 633 movdqa xmm6, xmm1 ; local 642 por xmm7, xmm6 ; local 654 movdqa xmm6, xmm1 ; local 663 por xmm7, xmm6 ; local 905 pxor xmm6, xmm6 ; local 923 paddd xmm6, xmm1 ; local 948 paddd xmm6, xmm2 local 955 psubd xmm6, xmm1 local 1186 movd xmm6, eax local 1214 paddd xmm6, xmm2 local 1217 pshufd xmm6, xmm6, 0 // duplicate the last ones local 1226 paddd xmm6, xmm4 local 1233 paddd xmm6, xmm4 local 1239 paddd xmm6, xmm4 local 1280 psrldq xmm6, 12 local [all...] |
H A D | recon_sse2.asm | 92 movdqa xmm6, xmm2 94 punpckhbw xmm6, xmm0 96 paddsw xmm6, XMMWORD PTR [rdx+48] 97 packuswb xmm2, xmm6 ; pack and unpack to saturate
|
/external/libyuv/files/source/ |
H A D | rotate.cc | 103 movq xmm6, qword ptr [eax] local 106 punpcklbw xmm6, xmm7 local 108 movdqa xmm7, xmm6 117 punpcklwd xmm4, xmm6 119 movdqa xmm6, xmm4 local 121 palignr xmm6, xmm6, 8 local 131 punpckldq xmm2, xmm6 132 movdqa xmm6, xmm2 local 133 palignr xmm6, xmm local 205 movdqa xmm6, [eax] local 211 punpcklbw xmm6, xmm7 local 228 movdqa xmm6, xmm5 local 231 movdqa xmm6, xmm5 // use xmm6 as temp register. local 233 punpckhwd xmm6, xmm7 local 237 movdqa xmm6, xmm0 local 239 punpckhdq xmm6, xmm4 local 241 movdqa xmm6, [esp] // restore xmm6 local [all...] |
H A D | row_win.cc | 85 movdqa xmm6, _kAddY16 local 102 paddb xmm0, xmm6 118 movdqa xmm6, _kAddY16 local 135 paddb xmm0, xmm6 151 movdqa xmm6, _kAddY16 local 168 paddb xmm0, xmm6 189 movdqa xmm6, _kARGBToV local 220 pmaddubsw xmm1, xmm6 // V 221 pmaddubsw xmm3, xmm6 253 movdqa xmm6, _kBGRATo local 317 movdqa xmm6, _kABGRToV local 377 movdqa xmm6, _kShuffleMaskBG24ToARGB local 416 movdqa xmm6, _kShuffleMaskRAWToARGB local [all...] |
H A D | convert.cc | 373 movq xmm6, QWORD PTR [ebx] ;src_u local 375 punpcklbw xmm6, xmm0 ;src_u, src_v mix local 376 ;movdqa xmm1, xmm6 local 377 ;movdqa xmm2, xmm6 local 378 ;movdqa xmm4, xmm6 382 punpcklbw xmm1, xmm6 ;in1, src_u, in1, src_v local 388 punpcklbw xmm2, xmm6 ;in2, src_u, in2, src_v local 392 punpckhbw xmm3, xmm6 ;in1, src_u, in1, src_v again local 397 punpckhbw xmm5, xmm6 ;src_u, in2, src_v again local 487 movq xmm6, QWOR local 489 punpcklbw xmm6, xmm0 ;src_u, src_v mix local 509 punpckhbw xmm6, xmm5 ;src_u, in2, src_v again local 510 movdqu XMMWORD PTR [edi+16], xmm6 ;write to out2 again local [all...] |
H A D | scale.cc | 503 movdqa xmm6, [ebp + edx] local 504 pavgb xmm4, xmm6 505 movdqa xmm6, [ebp + edx + 16] local 506 pavgb xmm5, xmm6 583 // xmm6 madd 1 601 movdqa xmm6, _madd11 local 618 pmaddubsw xmm0, xmm6 658 movdqa xmm6, _madd11 local 677 pmaddubsw xmm0, xmm6 717 movdqa xmm6, _shuf38 local 752 movdqa xmm6, _scaleac3 local 816 movdqa xmm6, _shufab2 local 917 movd xmm6, eax // xmm6 = y fraction local 918 punpcklwd xmm6, xmm6 local 919 pshufd xmm6, xmm6, 0 local 1086 movdqa xmm6, _madd11 local [all...] |
/external/openssl/crypto/aes/asm/ |
H A D | vpaes-x86_64.pl | 78 ## Preserves %xmm6 - %xmm8 so you get some local vectors 343 ## the high bits of %xmm6. 354 movdqa %xmm0, %xmm6 # save short part 356 movhlps %xmm4, %xmm6 # clobber low side with zeros 361 palignr \$8,%xmm6,%xmm0 379 ## %xmm6. The low side's rounds are the same as the 390 movdqa %xmm0, %xmm6 # save cur_lo in xmm6 398 # low round. swap xmm7 and xmm6 401 movdqa %xmm6, [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | x86_64-mont5.pl | 78 movaps %xmm6,(%rsp) 106 movq 16(%rax,%r10,8),%xmm6 # denoted by 7th argument 115 pand %xmm6,%xmm2 142 pand %xmm6,%xmm2 222 pand %xmm6,%xmm2 325 movaps (%rsi),%xmm6 360 movaps %xmm6,(%rsp) 389 movq 16(%rax,%r10,8),%xmm6 # denoted by 7th argument 398 pand %xmm6,%xmm2 424 pand %xmm6, [all...] |