/external/compiler-rt/lib/builtins/i386/ |
H A D | floatdisf.S | 21 movd 4(%esp), %xmm0 22 movd 8(%esp), %xmm1
|
H A D | floatdixf.S | 21 movd 4(%esp), %xmm0 22 movd 8(%esp), %xmm1
|
H A D | ashldi3.S | 21 movd 12(%esp), %xmm2 // Load count 23 movd 4(%esp), %xmm0 24 movd 8(%esp), %xmm1 30 movd %xmm0, %eax 32 movd %xmm0, %edx
|
H A D | lshrdi3.S | 21 movd 12(%esp), %xmm2 // Load count 23 movd 4(%esp), %xmm0 24 movd 8(%esp), %xmm1 30 movd %xmm0, %eax 32 movd %xmm0, %edx
|
H A D | ashrdi3.S | 14 movd 12(%esp), %xmm2 // Load count 17 movd 4(%esp), %xmm0 18 movd 8(%esp), %xmm1 40 1: movd %xmm0, %eax 42 movd %xmm0, %edx
|
H A D | floatundisf.S | 33 movd 8(%esp), %xmm1 34 movd 4(%esp), %xmm0 78 movd 8(%esp), %xmm1 79 movd 4(%esp), %xmm0 90 movd %edx, %xmm3 100 movd %xmm0, 4(%esp)
|
/external/swiftshader/third_party/LLVM/test/MC/X86/ |
H A D | x86_64-encoding.s | 85 // CHECK: movd %r8, %mm1 87 movd %r8, %mm1 label 89 // CHECK: movd %r8d, %mm1 91 movd %r8d, %mm1 label 93 // CHECK: movd %rdx, %mm1 95 movd %rdx, %mm1 label 97 // CHECK: movd %edx, %mm1 99 movd %edx, %mm1 label 101 // CHECK: movd %mm1, %r8 103 movd label 107 movd %mm1, %r8d label 111 movd %mm1, %rdx label 115 movd %mm1, %edx label [all...] |
/external/compiler-rt/lib/builtins/x86_64/ |
H A D | floatundidf.S | 39 movd %edi, %xmm0 // low 32 bits of a 43 movd %rdi, %xmm1
|
/external/llvm/test/MC/X86/ |
H A D | x86_64-encoding.s | 85 // CHECK: movd %r8, %mm1 87 movd %r8, %mm1 label 89 // CHECK: movd %r8d, %mm1 91 movd %r8d, %mm1 label 93 // CHECK: movd %rdx, %mm1 95 movd %rdx, %mm1 label 97 // CHECK: movd %edx, %mm1 99 movd %edx, %mm1 label 101 // CHECK: movd %mm1, %r8 103 movd label 107 movd %mm1, %r8d label 111 movd %mm1, %rdx label 115 movd %mm1, %edx label [all...] |
/external/boringssl/src/crypto/fipsmodule/bn/asm/ |
H A D | bn-586.pl | 47 &movd("mm0",&wparam(3)); # mm0 = w 52 &movd("mm3",&DWP(0,$r,"",0)); # mm3 = r[0] 54 &movd("mm2",&DWP(0,$a,"",0)); # mm2 = a[0] 56 &movd("mm4",&DWP(4,$a,"",0)); # mm4 = a[1] 58 &movd("mm6",&DWP(8,$a,"",0)); # mm6 = a[2] 60 &movd("mm7",&DWP(12,$a,"",0)); # mm7 = a[3] 63 &movd("mm3",&DWP(4,$r,"",0)); # mm3 = r[1] 65 &movd("mm5",&DWP(8,$r,"",0)); # mm5 = r[2] 67 &movd("mm4",&DWP(12,$r,"",0)); # mm4 = r[3] 69 &movd( [all...] |
H A D | x86-mont.pl | 149 &movd ($mask,"eax"); # mask 32 lower bits 158 &movd ($mul0,&DWP(0,$bp)); # bp[0] 159 &movd ($mul1,&DWP(0,$ap)); # ap[0] 160 &movd ($car1,&DWP(0,$np)); # np[0] 164 &movq ($acc0,$mul1); # I wish movd worked for 172 &movd ($acc1,&DWP(4,$np)); # np[1] 173 &movd ($acc0,&DWP(4,$ap)); # ap[1] 187 &movd ($acc1,&DWP(4,$np,$j,4)); # np[j+1] 189 &movd ($acc0,&DWP(4,$ap,$j,4)); # ap[j+1] 191 &movd ( [all...] |
/external/libvpx/libvpx/vpx_dsp/x86/ |
H A D | sad_sse3.asm | 196 movd [rcx], xmm0 202 movd [rcx+4], xmm0 208 movd [rcx+8], xmm0 234 movd [rcx], xmm0 240 movd [rcx+4], xmm0 246 movd [rcx+8], xmm0 275 movd [rcx+8], mm7 300 movd [rcx+8], mm7 315 movd mm0, DWORD PTR [src_ptr] 316 movd mm [all...] |
H A D | sad4d_sse2.asm | 17 movd m0, [srcq +%2] 19 movd m6, [ref1q+%3] 20 movd m4, [ref2q+%3] 21 movd m7, [ref3q+%3] 22 movd m5, [ref4q+%3] 23 movd m1, [srcq +%4] 24 movd m2, [ref1q+%5] 27 movd m1, [ref2q+%5] 28 movd m2, [ref3q+%5] 29 movd m [all...] |
H A D | vpx_convolve_copy_sse2.asm | 194 movd m0, [srcq] 195 movd m1, [srcq+src_strideq] 196 movd m2, [srcq+src_strideq*2] 197 movd m3, [srcq+r5q] 200 movd m4, [dstq] 201 movd m5, [dstq+dst_strideq] 202 movd m6, [dstq+dst_strideq*2] 203 movd m7, [dstq+r6q] 209 movd [dstq ], m0 210 movd [dst [all...] |
H A D | ssim_opt_x86_64.asm | 122 movd [rdi], xmm15; 124 movd [rdi], xmm14; 126 movd [rdi], xmm13; 128 movd [rdi], xmm12; 130 movd [rdi], xmm11; 203 movd [rdi], xmm15; 205 movd [rdi], xmm14; 207 movd [rdi], xmm13; 209 movd [rdi], xmm12; 211 movd [rd [all...] |
H A D | sad_sse2.asm | 80 movd eax, m0 123 movd eax, m0 169 movd eax, m0 213 movd eax, m0 233 movd m1, [refq] 234 movd m2, [refq+ref_strideq] 235 movd m3, [refq+ref_strideq*2] 236 movd m4, [refq+ref_stride3q] 244 movd m2, [srcq] 245 movd m [all...] |
H A D | intrapred_sse2.asm | 55 movd [dstq ], m3 57 movd [dstq+strideq ], m3 60 movd [dstq ], m3 62 movd [dstq+strideq ], m3 64 movd tempd, m0 113 movd m0, [leftq] ; abcd [byte] 125 movd [dstq ], m1 127 movd [dstq+strideq], m1 131 movd [dstq ], m1 132 movd [dst [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
H A D | iwalsh_sse2.asm | 58 movd xmm0, eax 80 movd eax, xmm5 81 movd ecx, xmm4 90 movd eax, xmm5 91 movd ecx, xmm4 101 movd eax, xmm5 102 movd ecx, xmm4 111 movd eax, xmm5 112 movd ecx, xmm4
|
H A D | idctllm_mmx.asm | 187 movd mm4, [rsi] 191 movd [rdx], mm0 193 movd mm4, [rsi+rax] 197 movd [rdx+rdi], mm1 199 movd mm4, [rsi+2*rax] 203 movd [rdx+rdi*2], mm2 208 movd mm4, [rsi+2*rax] 212 movd [rdx+rdi*2], mm5 236 movd mm5, arg(0) ;input_dc 251 movd mm [all...] |
/external/fec/ |
H A D | peak_mmx_assist.s | 40 2: movd %mm7,%eax 44 movd %mm7,%edx 51 movd %mm7,%edx 58 movd %mm7,%edx
|
H A D | peakval_mmx_assist.s | 40 2: movd %mm7,%eax 44 movd %mm7,%edx 51 movd %mm7,%edx 58 movd %mm7,%edx
|
/external/boringssl/linux-x86/crypto/fipsmodule/ |
H A D | x86-mont.S | 68 movd %eax,%mm7 74 movd (%edi),%mm4 75 movd (%esi),%mm5 76 movd (%ebp),%mm3 84 movd 4(%ebp),%mm1 85 movd 4(%esi),%mm0 97 movd 4(%ebp,%ecx,4),%mm1 99 movd 4(%esi,%ecx,4),%mm0 101 movd %mm3,28(%esp,%ecx,4) 113 movd [all...] |
/external/boringssl/mac-x86/crypto/fipsmodule/ |
H A D | x86-mont.S | 67 movd %eax,%mm7 73 movd (%edi),%mm4 74 movd (%esi),%mm5 75 movd (%ebp),%mm3 83 movd 4(%ebp),%mm1 84 movd 4(%esi),%mm0 96 movd 4(%ebp,%ecx,4),%mm1 98 movd 4(%esi,%ecx,4),%mm0 100 movd %mm3,28(%esp,%ecx,4) 112 movd [all...] |
/external/capstone/suite/MC/X86/ |
H A D | x86_64-encoding.s.cs | 21 0x49,0x0f,0x6e,0xc8 = movd %r8, %mm1 22 0x41,0x0f,0x6e,0xc8 = movd %r8d, %mm1 23 0x48,0x0f,0x6e,0xca = movd %rdx, %mm1 24 0x0f,0x6e,0xca = movd %edx, %mm1 25 0x49,0x0f,0x7e,0xc8 = movd %mm1, %r8 26 0x41,0x0f,0x7e,0xc8 = movd %mm1, %r8d 27 0x48,0x0f,0x7e,0xca = movd %mm1, %rdx 28 0x0f,0x7e,0xca = movd %mm1, %edx
|
/external/boringssl/win-x86/crypto/fipsmodule/ |
H A D | x86-mont.asm | 78 movd mm7,eax 84 movd mm4,DWORD [edi] 85 movd mm5,DWORD [esi] 86 movd mm3,DWORD [ebp] 94 movd mm1,DWORD [4+ebp] 95 movd mm0,DWORD [4+esi] 107 movd mm1,DWORD [4+ecx*4+ebp] 109 movd mm0,DWORD [4+ecx*4+esi] 111 movd DWORD [28+ecx*4+esp],mm3 123 movd DWOR [all...] |