/external/compiler-rt/lib/builtins/i386/ |
H A D | floatdisf.S | 21 movd 4(%esp), %xmm0 22 movd 8(%esp), %xmm1
|
H A D | floatdixf.S | 21 movd 4(%esp), %xmm0 22 movd 8(%esp), %xmm1
|
H A D | ashldi3.S | 21 movd 12(%esp), %xmm2 // Load count 23 movd 4(%esp), %xmm0 24 movd 8(%esp), %xmm1 30 movd %xmm0, %eax 32 movd %xmm0, %edx
|
H A D | lshrdi3.S | 21 movd 12(%esp), %xmm2 // Load count 23 movd 4(%esp), %xmm0 24 movd 8(%esp), %xmm1 30 movd %xmm0, %eax 32 movd %xmm0, %edx
|
H A D | ashrdi3.S | 14 movd 12(%esp), %xmm2 // Load count 17 movd 4(%esp), %xmm0 18 movd 8(%esp), %xmm1 40 1: movd %xmm0, %eax 42 movd %xmm0, %edx
|
H A D | floatundisf.S | 33 movd 8(%esp), %xmm1 34 movd 4(%esp), %xmm0 78 movd 8(%esp), %xmm1 79 movd 4(%esp), %xmm0 90 movd %edx, %xmm3 100 movd %xmm0, 4(%esp)
|
/external/compiler-rt/lib/builtins/x86_64/ |
H A D | floatundidf.S | 39 movd %edi, %xmm0 // low 32 bits of a 43 movd %rdi, %xmm1
|
/external/llvm/test/MC/X86/ |
H A D | x86_64-encoding.s | 85 // CHECK: movd %r8, %mm1 87 movd %r8, %mm1 label 89 // CHECK: movd %r8d, %mm1 91 movd %r8d, %mm1 label 93 // CHECK: movd %rdx, %mm1 95 movd %rdx, %mm1 label 97 // CHECK: movd %edx, %mm1 99 movd %edx, %mm1 label 101 // CHECK: movd %mm1, %r8 103 movd label 107 movd %mm1, %r8d label 111 movd %mm1, %rdx label 115 movd %mm1, %edx label [all...] |
/external/boringssl/src/crypto/bn/asm/ |
H A D | bn-586.pl | 42 &movd("mm0",&wparam(3)); # mm0 = w 47 &movd("mm3",&DWP(0,$r,"",0)); # mm3 = r[0] 49 &movd("mm2",&DWP(0,$a,"",0)); # mm2 = a[0] 51 &movd("mm4",&DWP(4,$a,"",0)); # mm4 = a[1] 53 &movd("mm6",&DWP(8,$a,"",0)); # mm6 = a[2] 55 &movd("mm7",&DWP(12,$a,"",0)); # mm7 = a[3] 58 &movd("mm3",&DWP(4,$r,"",0)); # mm3 = r[1] 60 &movd("mm5",&DWP(8,$r,"",0)); # mm5 = r[2] 62 &movd("mm4",&DWP(12,$r,"",0)); # mm4 = r[3] 64 &movd( [all...] |
H A D | x86-mont.pl | 122 &movd ($mask,"eax"); # mask 32 lower bits 131 &movd ($mul0,&DWP(0,$bp)); # bp[0] 132 &movd ($mul1,&DWP(0,$ap)); # ap[0] 133 &movd ($car1,&DWP(0,$np)); # np[0] 137 &movq ($acc0,$mul1); # I wish movd worked for 145 &movd ($acc1,&DWP(4,$np)); # np[1] 146 &movd ($acc0,&DWP(4,$ap)); # ap[1] 160 &movd ($acc1,&DWP(4,$np,$j,4)); # np[j+1] 162 &movd ($acc0,&DWP(4,$ap,$j,4)); # ap[j+1] 164 &movd ( [all...] |
/external/libvpx/libvpx/vpx_dsp/x86/ |
H A D | sad_sse3.asm | 194 movd [rcx], xmm0 200 movd [rcx+4], xmm0 206 movd [rcx+8], xmm0 232 movd [rcx], xmm0 238 movd [rcx+4], xmm0 244 movd [rcx+8], xmm0 273 movd [rcx+8], mm7 298 movd [rcx+8], mm7 313 movd mm0, DWORD PTR [src_ptr] 314 movd mm [all...] |
H A D | intrapred_sse2.asm | 31 movd m0, [aboveq] 38 movd [dstq ], m0 39 movd [dstq+strideq], m0 41 movd [dstq ], m0 42 movd [dstq+strideq], m0 52 movd m0, [leftq] 58 movd [dstq ], m0 59 movd [dstq+strideq], m0 61 movd [dstq ], m0 62 movd [dst [all...] |
H A D | ssim_opt_x86_64.asm | 119 movd [rdi], xmm15; 121 movd [rdi], xmm14; 123 movd [rdi], xmm13; 125 movd [rdi], xmm12; 127 movd [rdi], xmm11; 200 movd [rdi], xmm15; 202 movd [rdi], xmm14; 204 movd [rdi], xmm13; 206 movd [rdi], xmm12; 208 movd [rd [all...] |
H A D | sad4d_sse2.asm | 17 movd m0, [srcq +%2] 19 movd m6, [ref1q+%3] 20 movd m4, [ref2q+%3] 21 movd m7, [ref3q+%3] 22 movd m5, [ref4q+%3] 35 movd m1, [ref1q+%3] 36 movd m2, [ref2q+%3] 37 movd m3, [ref3q+%3] 38 movd m4, [ref4q+%3]
|
H A D | sad_sse2.asm | 80 movd eax, m0 123 movd eax, m0 169 movd eax, m0 213 movd eax, m0 233 movd m1, [refq] 234 movd m2, [refq+ref_strideq] 235 movd m3, [refq+ref_strideq*2] 236 movd m4, [refq+ref_stride3q] 244 movd m2, [srcq] 245 movd m [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
H A D | iwalsh_sse2.asm | 56 movd xmm0, eax 78 movd eax, xmm5 79 movd ecx, xmm4 88 movd eax, xmm5 89 movd ecx, xmm4 99 movd eax, xmm5 100 movd ecx, xmm4 109 movd eax, xmm5 110 movd ecx, xmm4
|
H A D | iwalsh_mmx.asm | 98 movd eax, mm1 99 movd ecx, mm0 108 movd eax, mm1 109 movd ecx, mm0 117 movd eax, mm6 118 movd ecx, mm5 127 movd eax, mm6 128 movd ecx, mm5
|
H A D | idctllm_mmx.asm | 186 movd mm4, [rsi] 190 movd [rdx], mm0 192 movd mm4, [rsi+rax] 196 movd [rdx+rdi], mm1 198 movd mm4, [rsi+2*rax] 202 movd [rdx+rdi*2], mm2 207 movd mm4, [rsi+2*rax] 211 movd [rdx+rdi*2], mm5 235 movd mm5, arg(0) ;input_dc 250 movd mm [all...] |
H A D | dequantize_mmx.asm | 216 movd mm4, [rdx] 220 movd [rdx], mm0 222 movd mm4, [rdx+rdi] 226 movd [rdx+rdi], mm1 228 movd mm4, [rdx+2*rdi] 232 movd [rdx+rdi*2], mm2 236 movd mm4, [rdx+2*rdi] 240 movd [rdx+rdi*2], mm5
|
/external/fec/ |
H A D | peak_mmx_assist.s | 40 2: movd %mm7,%eax 44 movd %mm7,%edx 51 movd %mm7,%edx 58 movd %mm7,%edx
|
H A D | peakval_mmx_assist.s | 40 2: movd %mm7,%eax 44 movd %mm7,%edx 51 movd %mm7,%edx 58 movd %mm7,%edx
|
H A D | sumsq_mmx_assist.s | 32 movd %mm0,%ebx 36 movd %mm0,%ebx 76 2: movd %mm2,%eax # even sum 78 movd %mm2,%edx # odd sum
|
/external/boringssl/linux-x86/crypto/bn/ |
H A D | x86-mont.S | 54 movd %eax,%mm7 60 movd (%edi),%mm4 61 movd (%esi),%mm5 62 movd (%ebp),%mm3 70 movd 4(%ebp),%mm1 71 movd 4(%esi),%mm0 83 movd 4(%ebp,%ecx,4),%mm1 85 movd 4(%esi,%ecx,4),%mm0 87 movd %mm3,28(%esp,%ecx,4) 99 movd [all...] |
/external/boringssl/mac-x86/crypto/bn/ |
H A D | x86-mont.S | 53 movd %eax,%mm7 59 movd (%edi),%mm4 60 movd (%esi),%mm5 61 movd (%ebp),%mm3 69 movd 4(%ebp),%mm1 70 movd 4(%esi),%mm0 82 movd 4(%ebp,%ecx,4),%mm1 84 movd 4(%esi,%ecx,4),%mm0 86 movd %mm3,28(%esp,%ecx,4) 98 movd [all...] |
/external/boringssl/win-x86/crypto/bn/ |
H A D | x86-mont.asm | 63 movd mm7,eax 69 movd mm4,DWORD [edi] 70 movd mm5,DWORD [esi] 71 movd mm3,DWORD [ebp] 79 movd mm1,DWORD [4+ebp] 80 movd mm0,DWORD [4+esi] 92 movd mm1,DWORD [4+ecx*4+ebp] 94 movd mm0,DWORD [4+ecx*4+esi] 96 movd DWORD [28+ecx*4+esp],mm3 108 movd DWOR [all...] |