/external/openssl/crypto/aes/asm/ |
H A D | vpaes-x86.pl | 161 &movdqa ("xmm7",&QWP($k_inv,$const)); 162 &movdqa ("xmm6",&QWP($k_s0F,$const)); 183 &movdqa ("xmm1","xmm6") 184 &movdqa ("xmm2",&QWP($k_ipt,$const)); 190 &movdqa ("xmm0",&QWP($k_ipt+16,$const)); 201 &movdqa ("xmm4",&QWP($k_sb1,$const)); # 4 : sb1u 204 &movdqa ("xmm0",&QWP($k_sb1+16,$const));# 0 : sb1t 207 &movdqa ("xmm5",&QWP($k_sb2,$const)); # 4 : sb2u 209 &movdqa ("xmm1",&QWP(-0x40,$base,$magic));# .Lk_mc_forward[] 210 &movdqa ("xmm [all...] |
H A D | vpaes-x86_64.pl | 87 movdqa %xmm9, %xmm1 88 movdqa .Lk_ipt(%rip), %xmm2 # iptlo 94 movdqa .Lk_ipt+16(%rip), %xmm0 # ipthi 105 movdqa %xmm13, %xmm4 # 4 : sb1u 108 movdqa %xmm12, %xmm0 # 0 : sb1t 111 movdqa %xmm15, %xmm5 # 4 : sb2u 113 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 114 movdqa %xmm14, %xmm2 # 2 : sb2t 117 movdqa (%r11,%r10), %xmm4 # .Lk_mc_backward[] 118 movdqa [all...] |
H A D | bsaes-x86_64.pl | 57 # SIMD "domain" (movaps instead of movdqa, xorps instead of 229 movdqa $y0, $t0 244 movdqa $y0, $t0 260 movdqa $y0, $t0 261 movdqa $y2, $t1 283 movdqa @x[0], @t[0] 284 movdqa @x[1], @t[1] 301 movdqa @x[4], @t[0] 302 movdqa @x[5], @t[1] 329 movdqa [all...] |
/external/libvpx/vp8/encoder/x86/ |
H A D | subtract_sse2.asm | 95 movdqa xmm0, XMMWORD PTR [rsi] ; src 96 movdqa xmm1, XMMWORD PTR [rax] ; pred 98 movdqa xmm2, xmm0 105 movdqa xmm2, xmm0 106 movdqa xmm3, xmm1 110 movdqa XMMWORD PTR [rdi], xmm0 111 movdqa XMMWORD PTR [rdi +16], xmm2 113 movdqa xmm4, XMMWORD PTR [rsi + rdx] 114 movdqa xmm5, XMMWORD PTR [rax + 16] 116 movdqa xmm [all...] |
H A D | temporal_filter_apply_sse2.asm | 47 movdqa [rsp + strength], xmm6 ; where strength is used, all 16 bytes are read 54 movdqa xmm5, [GLOBAL(_const_top_bit)] 56 movdqa [rsp + rounding_bit], xmm5 67 movdqa [rsp + filter_weight], xmm0 87 movdqa xmm0, [rsi] ; src (frame1) 89 movdqa xmm1, xmm0 94 movdqa xmm2, [rdx] ; predictor (frame2) 95 movdqa xmm3, xmm2 121 movdqa xmm3, [GLOBAL(_const_16w)] 122 movdqa xmm [all...] |
H A D | dct_sse2.asm | 76 movdqa xmm2, xmm0 79 movdqa xmm1, xmm0 85 movdqa xmm3, xmm0 91 movdqa xmm1, xmm0 94 movdqa xmm4, xmm3 109 movdqa xmm2, xmm0 113 movdqa xmm3, xmm0 116 movdqa xmm2, xmm0 120 movdqa xmm5, XMMWORD PTR[GLOBAL(_7)] 122 movdqa xmm [all...] |
H A D | quantize_ssse3.asm | 33 movdqa xmm0, [rdx] 34 movdqa xmm4, [rdx + 16] 36 movdqa xmm2, [rdi] ;round lo 37 movdqa xmm3, [rdi + 16] ;round hi 39 movdqa xmm1, xmm0 40 movdqa xmm5, xmm4 63 movdqa [rdi], xmm1 64 movdqa [rdi + 16], xmm5 66 movdqa xmm2, [rcx] 67 movdqa xmm [all...] |
H A D | quantize_sse2.asm | 62 movdqa xmm0, [rdx] 63 movdqa xmm4, [rdx + 16] 69 movdqa xmm1, xmm0 70 movdqa xmm5, xmm4 84 movdqa xmm2, [rcx] 85 movdqa xmm3, [rcx + 16] 95 movdqa [rsp + abs_minus_zbin], xmm1 96 movdqa [rsp + abs_minus_zbin + 16], xmm5 102 movdqa xmm2, [rdx] 103 movdqa xmm [all...] |
H A D | fwalsh_sse2.asm | 40 movdqa xmm1, xmm0 44 movdqa xmm2, xmm0 51 movdqa xmm1, xmm0 61 movdqa xmm2, xmm0 77 movdqa xmm2, xmm0 80 movdqa xmm3, xmm1 89 movdqa xmm0, xmm4 92 movdqa xmm1, xmm6 96 movdqa xmm2, xmm0 99 movdqa xmm [all...] |
H A D | sad_ssse3.asm | 16 movdqa xmm0, XMMWORD PTR [rsi] 25 movdqa xmm0, XMMWORD PTR [rsi] 38 movdqa xmm0, XMMWORD PTR [rsi+rax] 57 movdqa xmm0, XMMWORD PTR [rsi] 58 movdqa xmm4, XMMWORD PTR [rdi] 59 movdqa xmm7, XMMWORD PTR [rdi+16] 61 movdqa xmm5, xmm7 64 movdqa xmm6, xmm7 73 movdqa xmm0, XMMWORD PTR [rsi] 74 movdqa xmm [all...] |
H A D | encodeopt.asm | 27 movdqa xmm3, [rsi] 29 movdqa xmm4, [rdi] 30 movdqa xmm5, [rsi+16] 32 movdqa xmm6, [rdi+16] 42 movdqa xmm0, xmm3 48 movdqa xmm3, xmm0 229 movdqa xmm3, [rsi] 230 movdqa xmm4, [rdi] 232 movdqa xmm5, [rsi+16] 233 movdqa xmm [all...] |
H A D | sad_sse4.asm | 16 movdqa xmm0, XMMWORD PTR [rsi] 23 movdqa xmm2, xmm1 29 movdqa xmm4, xmm3 37 movdqa xmm0, XMMWORD PTR [rsi] 44 movdqa xmm2, xmm5 50 movdqa xmm4, xmm3 60 movdqa xmm0, XMMWORD PTR [rsi + rax] 70 movdqa xmm2, xmm5 75 movdqa xmm4, xmm3 93 movdqa xmm [all...] |
/external/libvpx/vp8/common/x86/ |
H A D | idctllm_sse2.asm | 124 movdqa xmm0, [rax] 125 movdqa xmm2, [rax+16] 126 movdqa xmm1, [rax+32] 127 movdqa xmm3, [rax+48] 130 movdqa [rax], xmm7 131 movdqa [rax+16], xmm7 132 movdqa [rax+32], xmm7 133 movdqa [rax+48], xmm7 142 movdqa xmm4, xmm0 149 movdqa xmm [all...] |
H A D | iwalsh_sse2.asm | 29 movdqa xmm0, [rsi + 0] ;ip[4] ip[0] 30 movdqa xmm1, [rsi + 16] ;ip[12] ip[8] 36 movdqa xmm3, xmm0 ;ip[4] ip[0] 41 movdqa xmm4, xmm0 46 movdqa xmm1, xmm4 ;c1 b1 59 movdqa xmm3, xmm4 ; 13 12 11 10 03 02 01 00 62 movdqa xmm1, xmm4 ; 23 03 22 02 21 01 20 00 67 movdqa xmm3, xmm4 ;ip[4] ip[0] 74 movdqa xmm5, xmm4 78 movdqa xmm [all...] |
H A D | loopfilter_sse2.asm | 19 movdqa xmm2, [rdi+2*rax] ; q3 20 movdqa xmm1, [rsi+2*rax] ; q2 21 movdqa xmm4, [rsi+rax] ; q1 22 movdqa xmm5, [rsi] ; q0 38 movdqa XMMWORD PTR [rsp], xmm1 ; store q2 39 movdqa XMMWORD PTR [rsp + 16], xmm4 ; store q1 42 movdqa xmm6, xmm1 ; q2 43 movdqa xmm3, xmm4 ; q1 54 movdqa xmm0, xmm5 ; q0 61 movdqa t [all...] |
H A D | recon_sse2.asm | 81 movdqa xmm1, XMMWORD PTR [rsi] 82 movdqa xmm5, xmm1 88 movdqa XMMWORD PTR [rdi], xmm1 91 movdqa xmm2, XMMWORD PTR [rsi+16] 92 movdqa xmm6, xmm2 98 movdqa XMMWORD PTR [rdi+rax], xmm2 101 movdqa xmm3, XMMWORD PTR [rsi+32] 102 movdqa xmm7, xmm3 108 movdqa XMMWORD PTR [rdi+rax*2], xmm3 111 movdqa xmm [all...] |
H A D | subpixel_sse2.asm | 67 movdqa xmm4, xmm1 68 movdqa xmm5, xmm1 70 movdqa xmm6, xmm1 71 movdqa xmm7, xmm1 117 movdqa XMMWORD Ptr [rdi], xmm4 192 movdqa xmm4, xmm1 193 movdqa xmm5, xmm1 195 movdqa xmm6, xmm1 196 movdqa xmm7, xmm1 241 movdqa XMMWOR [all...] |
H A D | subpixel_ssse3.asm | 51 movdqa xmm7, [GLOBAL(rd)] 60 movdqa xmm4, XMMWORD PTR [rax] ;k0_k5 61 movdqa xmm5, XMMWORD PTR [rax+256] ;k2_k4 62 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 79 movdqa xmm1, xmm0 82 movdqa xmm2, xmm1 115 movdqa xmm5, XMMWORD PTR [rax+256] ;k2_k4 116 movdqa xmm6, XMMWORD PTR [rax+128] ;k1_k3 118 movdqa xmm3, XMMWORD PTR [GLOBAL(shuf2bfrom1)] 119 movdqa xmm [all...] |
/external/libyuv/files/source/ |
H A D | row_win.cc | 84 movdqa xmm7, _kARGBToY 85 movdqa xmm6, _kAddY16 88 movdqa xmm0, [eax] 89 movdqa xmm1, [eax + 16] 90 movdqa xmm2, [eax + 32] 91 movdqa xmm3, [eax + 48] 103 movdqa [edx], xmm0 local 117 movdqa xmm7, _kBGRAToY 118 movdqa xmm6, _kAddY16 121 movdqa xmm 136 movdqa [edx], xmm0 local 169 movdqa [edx], xmm0 local 390 movdqa [edx + 32], xmm2 local 393 movdqa [edx], xmm0 local 397 movdqa [edx + 16], xmm1 local 399 movdqa [edx + 48], xmm3 local 429 movdqa [edx + 32], xmm2 local 432 movdqa [edx], xmm0 local 436 movdqa [edx + 16], xmm1 local 438 movdqa [edx + 48], xmm3 local [all...] |
H A D | rotate.cc | 91 movdqa xmm1, xmm0 96 movdqa xmm3, xmm2 102 movdqa xmm5, xmm4 108 movdqa xmm7, xmm6 113 movdqa xmm2, xmm0 114 movdqa xmm3, xmm1 119 movdqa xmm6, xmm4 120 movdqa xmm7, xmm5 127 movdqa xmm4, xmm0 132 movdqa xmm 208 movdqa [esp], xmm5 // backup xmm5 local 230 movdqa [esp], xmm6 // backup xmm6 local 874 movdqa [edx], xmm0 local [all...] |
H A D | scale.cc | 281 movdqa xmm0, [eax] 282 movdqa xmm1, [eax + 16] 287 movdqa [edx], xmm0 local 310 movdqa xmm0, [eax] 311 movdqa xmm1, [eax + 16] 312 movdqa xmm2, [eax + esi] 313 movdqa xmm3, [eax + esi + 16] 318 movdqa xmm2, xmm0 // average columns (32 to 16 pixels) 320 movdqa xmm3, xmm1 328 movdqa [ed local 885 movdqa [edi], xmm2 local 886 movdqa [edi + 16], xmm3 local 946 movdqa [edi], xmm0 local 960 movdqa [edi], xmm0 local 976 movdqa [edi], xmm0 local 1028 movdqa [edi], xmm0 local 1042 movdqa [edi], xmm0 local 1058 movdqa [edi], xmm0 local [all...] |
H A D | planar_functions.cc | 76 movdqa xmm0, [eax] 77 movdqa xmm1, [eax + 16] 79 movdqa xmm2, xmm0 80 movdqa xmm3, xmm1 84 movdqa [edx], xmm0 89 movdqa [edi], xmm2 107 "movdqa (%0),%%xmm0\n" 108 "movdqa 0x10(%0),%%xmm1\n" 110 "movdqa %%xmm0,%%xmm2\n" 111 "movdqa 525 movdqa [edx], xmm2 local 675 movdqa [edx], xmm0 local 742 movdqa [edx], xmm0 local 1222 movdqa [edx], xmm0 local 1223 movdqa [edx + 16], xmm1 local 1245 movdqa [edx], xmm0 local 1267 movdqa [edx], xmm0 local [all...] |
/external/libffi/src/x86/ |
H A D | unix64.S | 194 movdqa 48(%r10), %xmm0 195 movdqa 64(%r10), %xmm1 196 movdqa 80(%r10), %xmm2 197 movdqa 96(%r10), %xmm3 198 movdqa 112(%r10), %xmm4 199 movdqa 128(%r10), %xmm5 200 movdqa 144(%r10), %xmm6 201 movdqa 160(%r10), %xmm7 318 movdqa %xmm0, 48(%rsp) 319 movdqa [all...] |
H A D | darwin64.S | 190 movdqa 48(%r10), %xmm0 191 movdqa 64(%r10), %xmm1 192 movdqa 80(%r10), %xmm2 193 movdqa 96(%r10), %xmm3 194 movdqa 112(%r10), %xmm4 195 movdqa 128(%r10), %xmm5 196 movdqa 144(%r10), %xmm6 197 movdqa 160(%r10), %xmm7 307 movdqa %xmm0, 48(%rsp) 308 movdqa [all...] |
/external/openssl/crypto/sha/asm/ |
H A D | sha1-586.pl | 432 &movdqa (@X[3],&QWP(0,$tmp1)); # K_00_19 433 &movdqa (@X[4],&QWP(16,$tmp1)); # K_20_39 434 &movdqa (@X[5],&QWP(32,$tmp1)); # K_40_59 435 &movdqa (@X[6],&QWP(48,$tmp1)); # K_60_79 436 &movdqa (@X[2],&QWP(64,$tmp1)); # pbswap mask 467 &movdqa (&QWP(112+0,"esp"),@X[4]); # copy constants 468 &movdqa (&QWP(112+16,"esp"),@X[5]); 469 &movdqa (&QWP(112+32,"esp"),@X[6]); 471 &movdqa (&QWP(112+48,"esp"),@X[3]); 473 &movdqa ( [all...] |