/external/openssl/crypto/aes/asm/ |
H A D | bsaes-x86_64.pl | 58 # pxor, etc.). While it resulted in nominal 4% improvement on 133 pxor @b[6], @b[5] 134 pxor @b[1], @b[2] 135 pxor @b[0], @b[3] 136 pxor @b[2], @b[6] 137 pxor @b[0], @b[5] 139 pxor @b[3], @b[6] 140 pxor @b[7], @b[3] 141 pxor @b[5], @b[7] 142 pxor [all...] |
H A D | vpaes-x86.pl | 192 &pxor ("xmm2","xmm5"); 193 &pxor ("xmm0","xmm2"); 203 &pxor ("xmm4","xmm5"); # 4 = sb1u + k 206 &pxor ("xmm0","xmm4"); # 0 = A 212 &pxor ("xmm2","xmm5"); # 2 = 2A 217 &pxor ("xmm0","xmm2"); # 0 = 2A+B 220 &pxor ("xmm3","xmm0"); # 3 = 2A+B+D 223 &pxor ("xmm0","xmm3"); # 0 = 2A+3B+C+D 234 &pxor ("xmm0","xmm1"); # 0 = j 237 &pxor ("xmm [all...] |
H A D | vpaes-x86.s | 82 pxor %xmm5,%xmm2 83 pxor %xmm2,%xmm0 91 pxor %xmm5,%xmm4 94 pxor %xmm4,%xmm0 100 pxor %xmm5,%xmm2 105 pxor %xmm2,%xmm0 108 pxor %xmm0,%xmm3 111 pxor %xmm3,%xmm0 120 pxor %xmm1,%xmm0 123 pxor [all...] |
H A D | vpaes-x86_64.pl | 96 pxor %xmm5, %xmm2 97 pxor %xmm2, %xmm0 107 pxor %xmm5, %xmm4 # 4 = sb1u + k 110 pxor %xmm4, %xmm0 # 0 = A 116 pxor %xmm5, %xmm2 # 2 = 2A 121 pxor %xmm2, %xmm0 # 0 = 2A+B 124 pxor %xmm0, %xmm3 # 3 = 2A+B+D 127 pxor %xmm3, %xmm0 # 0 = 2A+3B+C+D 138 pxor %xmm1, %xmm0 # 0 = j 141 pxor [all...] |
H A D | aesni-x86.s | 59 pxor %xmm0,%xmm3 60 pxor %xmm0,%xmm4 90 pxor %xmm0,%xmm3 91 pxor %xmm0,%xmm4 121 pxor %xmm0,%xmm3 122 pxor %xmm0,%xmm4 123 pxor %xmm0,%xmm5 157 pxor %xmm0,%xmm3 158 pxor %xmm0,%xmm4 159 pxor [all...] |
H A D | aesni-x86.pl | 217 &pxor ($inout1,$rndkey0); 218 &pxor ($inout2,$rndkey0); 256 &pxor ($inout1,$rndkey0); 257 &pxor ($inout2,$rndkey0); 258 &pxor ($inout3,$rndkey0); 298 &pxor ($inout1,$rndkey0); # pxor does better here 300 &pxor ($inout2,$rndkey0); 302 &pxor ($inout3,$rndkey0); 305 &pxor ( [all...] |
H A D | aesni-x86_64.pl | 382 pxor $rndkey0,$inout1 384 pxor $rndkey0,$inout2 386 pxor $rndkey0,$inout3 388 pxor $rndkey0,$inout4 390 pxor $rndkey0,$inout5 448 pxor $rndkey0,$inout2 450 pxor $rndkey0,$inout3 452 pxor $rndkey0,$inout4 454 pxor $rndkey0,$inout5 457 pxor [all...] |
H A D | aes-586.pl | 404 &pxor ("mm0","mm3"); 693 &pxor ("mm0",&QWP(0,$key)); # 7, 6, 5, 4, 3, 2, 1, 0 694 &pxor ("mm4",&QWP(8,$key)); # 15,14,13,12,11,10, 9, 8 723 &pxor ("mm3","mm3"); &pxor ("mm7","mm7"); 729 &pxor ("mm0","mm3"); &pxor ("mm4","mm7"); # = r2 731 &pxor ("mm1","mm0"); &pxor ("mm5","mm4"); # r0^r2 732 &pxor ("mm [all...] |
/external/openssl/crypto/modes/asm/ |
H A D | ghash-x86.s | 220 pxor 8(%esi,%edx,1),%mm0 224 pxor (%esi,%edx,1),%mm1 227 pxor %mm2,%mm0 232 pxor 8(%esi,%ecx,1),%mm0 235 pxor (%eax,%ebp,8),%mm1 237 pxor (%esi,%ecx,1),%mm1 239 pxor %mm2,%mm0 243 pxor 8(%esi,%edx,1),%mm0 246 pxor (%eax,%ebx,8),%mm1 248 pxor ( [all...] |
H A D | ghash-x86.pl | 380 &pxor ($Zlo,&QWP(8,$Htbl,$nix)); 384 &pxor ($Zhi,&QWP(0,$rem_4bit,$rem[1],8)) if ($cnt<28); 386 &pxor ($Zhi,&QWP(0,$Htbl,$nix)); 389 &pxor ($Zlo,$tmp); 526 &pxor ($Zlo,&QWP(8,$Htbl,$nhi)); 529 &pxor ($Zhi,&QWP(0,$rem_4bit,$rem,8)); 532 &pxor ($Zhi,&QWP(0,$Htbl,$nhi)); 534 &pxor ($Zlo,$tmp); 543 &pxor ($Zlo,&QWP(8,$Htbl,$nlo)); 545 &pxor ( [all...] |
H A D | ghash-x86_64.pl | 363 pxor $Xi,$T1 # 364 pxor $Hkey,$T2 370 pxor $Xi,$T1 # 371 pxor $Xhi,$T1 # 376 pxor $T1,$Xhi 377 pxor $T2,$Xi # 388 pxor $T1,$Xi # 390 pxor $T1,$Xi # 395 pxor $T1,$Xi 396 pxor [all...] |
/external/valgrind/main/memcheck/tests/amd64/ |
H A D | xor-undef-amd64.stdout.exp | 14 Complain sse2 pxor 16 No complain sse2 pxor
|
/external/openssl/crypto/ |
H A D | x86_64cpuid.pl | 216 pxor %xmm0,%xmm0 217 pxor %xmm1,%xmm1 218 pxor %xmm2,%xmm2 219 pxor %xmm3,%xmm3 220 pxor %xmm4,%xmm4 221 pxor %xmm5,%xmm5 222 pxor %xmm6,%xmm6 223 pxor %xmm7,%xmm7 224 pxor %xmm8,%xmm8 225 pxor [all...] |
H A D | x86cpuid.pl | 236 &pxor ("xmm0","xmm0"); 237 &pxor ("xmm1","xmm1"); 238 &pxor ("xmm2","xmm2"); 239 &pxor ("xmm3","xmm3"); 240 &pxor ("xmm4","xmm4"); 241 &pxor ("xmm5","xmm5"); 242 &pxor ("xmm6","xmm6"); 243 &pxor ("xmm7","xmm7");
|
/external/openssl/crypto/sha/asm/ |
H A D | sha512-586.pl | 84 &pxor ("mm3","mm2"); 86 &pxor ("mm3","mm1"); 88 &pxor ("mm3","mm2"); 90 &pxor ("mm3","mm1"); 92 &pxor ("mm3","mm2"); # T1=Sigma1_512(e) 94 &pxor ("mm5","mm6"); # f^=g 98 &pxor ("mm5","mm6"); # f^=g 112 &pxor ("mm7","mm6"); 114 &pxor ("mm7","mm5"); 116 &pxor ("mm [all...] |
/external/libvpx/vp8/encoder/x86/ |
H A D | quantize_mmx.asm | 37 pxor mm3, mm0 55 pxor mm3, mm0 80 pxor mm7, mm4 95 pxor mm7, mm4 121 pxor mm7, mm4 136 pxor mm7, mm4 162 pxor mm7, mm4 177 pxor mm7, mm4 198 pxor mm5, mm5 199 pxor mm [all...] |
H A D | subtract_sse2.asm | 32 pxor mm7, mm7 101 pxor xmm1, [GLOBAL(t80)] ;convert to signed values 102 pxor xmm2, [GLOBAL(t80)] 119 pxor xmm5, [GLOBAL(t80)] ;convert to signed values 120 pxor xmm6, [GLOBAL(t80)] 177 pxor xmm1, [GLOBAL(t80)] ;convert to signed values 178 pxor xmm2, [GLOBAL(t80)] 198 pxor xmm1, [GLOBAL(t80)] ;convert to signed values 199 pxor xmm2, [GLOBAL(t80)] 221 pxor xmm [all...] |
H A D | quantize_sse2.asm | 77 pxor xmm1, xmm0 78 pxor xmm5, xmm4 123 pxor xmm6, xmm6 183 pxor xmm2, xmm0 184 pxor xmm3, xmm4 208 pxor xmm2, xmm6 209 pxor xmm3, xmm6 276 pxor xmm1, xmm0 277 pxor xmm5, xmm4 293 pxor xmm [all...] |
H A D | variance_impl_sse2.asm | 34 pxor xmm4, xmm4 122 pxor xmm0, xmm0 ; clear xmm0 for unpack 123 pxor xmm7, xmm7 ; clear xmm7 for accumulating diffs 125 pxor xmm6, xmm6 ; clear xmm6 for accumulating sse 166 pxor xmm6, xmm6 168 pxor xmm5, xmm5 238 pxor xmm0, xmm0 ; clear xmm0 for unpack 239 pxor xmm7, xmm7 ; clear xmm7 for accumulating diffs 241 pxor xmm6, xmm6 ; clear xmm6 for accumulating sse 277 pxor xmm [all...] |
H A D | ssim_opt.asm | 78 pxor xmm0, xmm0 79 pxor xmm15,xmm15 ;sum_s 80 pxor xmm14,xmm14 ;sum_r 81 pxor xmm13,xmm13 ;sum_sq_s 82 pxor xmm12,xmm12 ;sum_sq_r 83 pxor xmm11,xmm11 ;sum_sxr 166 pxor xmm0, xmm0 167 pxor xmm15,xmm15 ;sum_s 168 pxor xmm14,xmm14 ;sum_r 169 pxor xmm1 [all...] |
H A D | encodeopt.asm | 41 pxor xmm7, xmm7 74 pxor mm7, mm7 83 pxor mm1, mm1 ; from movd mm1, dc ; dc =0 140 pxor mm7, mm7 143 pxor mm2, mm2 217 pxor xmm7, xmm7 220 pxor xmm2, xmm2 288 pxor mm7, mm7 345 pxor xmm7, xmm7 363 pxor xmm [all...] |
H A D | variance_impl_ssse3.asm | 43 pxor xmm6, xmm6 44 pxor xmm7, xmm7 122 pxor xmm4, xmm4 188 pxor xmm4, xmm4 222 pxor xmm0, xmm0 257 pxor xmm0, xmm0 279 pxor xmm4, xmm4 306 pxor xmm0, xmm0 307 pxor xmm1, xmm1 308 pxor xmm [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | x86-gf2m.s | 17 pxor %mm5,%mm5 18 pxor %mm4,%mm4 52 pxor %mm2,%mm0 58 pxor %mm1,%mm0 64 pxor %mm2,%mm0 70 pxor %mm1,%mm0 76 pxor %mm2,%mm0 82 pxor %mm1,%mm0 88 pxor %mm2,%mm0 94 pxor [all...] |
H A D | x86-gf2m.pl | 68 &pxor ($B31,$B31); 69 &pxor ($B30,$B30); 104 &pxor ($R,@T[1]); 109 &pxor ($R,$B30); 111 &pxor ($R,@T[1]); 114 &pxor ($R,$B31); 117 &pxor ($R,@T[0]); 242 &pxor ($R,"mm7"); 244 &pxor ($R,"mm6"); # (a0+a1)�(b0+b1)-a1�b1-a0�b0 251 &pxor ( [all...] |
/external/libvpx/vp8/common/x86/ |
H A D | loopfilter_mmx.asm | 132 pxor mm5, mm5 147 pxor mm4, mm5 153 pxor mm2, [GLOBAL(t80)] ; p1 offset to convert to signed values 154 pxor mm7, [GLOBAL(t80)] ; q1 offset to convert to signed values 157 pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values 158 pxor mm0, [GLOBAL(t80)] ; offset to convert to signed values 169 pxor mm0, mm0 ; 170 pxor mm5, mm5 178 pxor mm0, mm0 ; 0 182 pxor mm [all...] |