/external/llvm/test/MC/X86/ |
H A D | x86-32-coverage.s | 524 // CHECK: movd %ecx, %xmm5 525 movd %ecx,%xmm5 527 // CHECK: movd 3735928559(%ebx,%ecx,8), %xmm5 528 movd 0xdeadbeef(%ebx,%ecx,8),%xmm5 530 // CHECK: movd %xmm5, %ecx 531 movd %xmm5,%ecx 533 // CHECK: movd %xmm5, 3735928559(%ebx,%ecx,8) 534 movd %xmm5,0xdeadbeef(%ebx,%ecx,8) 545 // CHECK: movq %xmm5, %xmm5 [all...] |
H A D | x86-32-fma3.s | 3 // CHECK: vfmadd132pd %xmm2, %xmm5, %xmm1 5 vfmadd132pd %xmm2, %xmm5, %xmm1 7 // CHECK: vfmadd132pd (%eax), %xmm5, %xmm1 9 vfmadd132pd (%eax), %xmm5, %xmm1 11 // CHECK: vfmadd132ps %xmm2, %xmm5, %xmm1 13 vfmadd132ps %xmm2, %xmm5, %xmm1 15 // CHECK: vfmadd132ps (%eax), %xmm5, %xmm1 17 vfmadd132ps (%eax), %xmm5, %xmm1 19 // CHECK: vfmadd213pd %xmm2, %xmm5, %xmm1 21 vfmadd213pd %xmm2, %xmm5, [all...] |
H A D | x86_64-xop-encoding.s | 56 // CHECK: vphadduwd (%rax), %xmm5 58 vphadduwd (%rax), %xmm5 75 // CHECK: vphaddubw %xmm5, %xmm0 77 vphaddubw %xmm5, %xmm0 88 // CHECK: vphaddubd (%rax), %xmm5 90 vphaddubd (%rax), %xmm5 91 // CHECK: vphaddubd %xmm5, %xmm7 93 vphaddubd %xmm5, %xmm7 99 // CHECK: vphadddq %xmm4, %xmm5 101 vphadddq %xmm4, %xmm5 [all...] |
H A D | x86-32-avx.s | 35 // CHECK: vaddss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 37 vaddss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 39 // CHECK: vsubss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 41 vsubss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 43 // CHECK: vmulss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 45 vmulss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 47 // CHECK: vdivss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 49 vdivss 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 51 // CHECK: vaddsd 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 53 vaddsd 3735928559(%ebx,%ecx,8), %xmm2, %xmm5 [all...] |
/external/libvpx/libvpx/vp9/encoder/x86/ |
H A D | vp9_subpel_variance_impl_sse2.asm | 45 movdqu xmm5, XMMWORD PTR [rsi] 47 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) horizontal line 1 56 pavgb xmm5, xmm1 ; xmm = vertical average of the above 58 movdqa xmm4, xmm5 59 punpcklbw xmm5, xmm0 ; xmm5 = words of above 64 psubw xmm5, xmm3 ; xmm5 -= xmm3 70 paddw xmm6, xmm5 ; xmm [all...] |
H A D | vp9_sad_sse4.asm | 38 movq xmm5, MMWORD PTR [rdi] 41 punpcklqdq xmm5, xmm3 44 movdqa xmm2, xmm5 45 mpsadbw xmm5, xmm0, 0x0 54 paddw xmm5, xmm2 55 paddw xmm5, xmm3 56 paddw xmm5, xmm4 58 paddw xmm1, xmm5 61 movq xmm5, MMWORD PTR [rdi+ rdx] 64 punpcklqdq xmm5, xmm [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/x86/ |
H A D | idctllm_sse2.asm | 34 movd xmm5, [rdx] 37 pinsrw xmm5, [rdx], 4 39 pmullw xmm4, xmm5 41 ; Zero out xmm5, for use unpacking 42 pxor xmm5, xmm5 45 movd [rax], xmm5 46 movd [rax+32], xmm5 64 punpcklbw xmm0, xmm5 65 punpcklbw xmm1, xmm5 [all...] |
H A D | sad_sse4.asm | 38 movq xmm5, MMWORD PTR [rdi] 41 punpcklqdq xmm5, xmm3 44 movdqa xmm2, xmm5 45 mpsadbw xmm5, xmm0, 0x0 54 paddw xmm5, xmm2 55 paddw xmm5, xmm3 56 paddw xmm5, xmm4 58 paddw xmm1, xmm5 61 movq xmm5, MMWORD PTR [rdi+ rdx] 64 punpcklqdq xmm5, xmm [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
H A D | idctllm_sse2.asm | 34 movd xmm5, [rdx] 37 pinsrw xmm5, [rdx], 4 39 pmullw xmm4, xmm5 41 ; Zero out xmm5, for use unpacking 42 pxor xmm5, xmm5 45 movd [rax], xmm5 46 movd [rax+32], xmm5 64 punpcklbw xmm0, xmm5 65 punpcklbw xmm1, xmm5 [all...] |
H A D | sad_sse4.asm | 38 movq xmm5, MMWORD PTR [rdi] 41 punpcklqdq xmm5, xmm3 44 movdqa xmm2, xmm5 45 mpsadbw xmm5, xmm0, 0x0 54 paddw xmm5, xmm2 55 paddw xmm5, xmm3 56 paddw xmm5, xmm4 58 paddw xmm1, xmm5 61 movq xmm5, MMWORD PTR [rdi+ rdx] 64 punpcklqdq xmm5, xmm [all...] |
/external/chromium_org/third_party/yasm/source/patched-yasm/modules/arch/x86/tests/ |
H A D | sse3.asm | 2 addsubpd xmm5, xmm7 4 addsubps xmm1, xmm5 13 hsubpd xmm5, xmm3 24 movsldup xmm5, dqword [eax+ebx]
|
H A D | simd-1.asm | 3 movntps dqword [0], xmm5
|
/external/chromium_org/third_party/libjpeg_turbo/simd/ |
H A D | jiss2fst-64.asm | 143 pshufd xmm5,xmm0,0xAA ; xmm5=col2=(02 02 02 02 02 02 02 02) 168 movdqa xmm5,xmm1 172 paddw xmm5,xmm3 ; xmm5=tmp13 176 psubw xmm1,xmm5 ; xmm1=tmp12 180 psubw xmm4,xmm5 ; xmm4=tmp3 182 paddw xmm6,xmm5 ; xmm6=tmp0 194 movdqa xmm5, XMMWORD [XMMBLOCK(5,0,rsi,SIZEOF_JCOEF)] 196 pmullw xmm5, XMMWOR [all...] |
H A D | jiss2fst.asm | 149 pshufd xmm5,xmm0,0xAA ; xmm5=col2=(02 02 02 02 02 02 02 02) 175 movdqa xmm5,xmm1 179 paddw xmm5,xmm3 ; xmm5=tmp13 183 psubw xmm1,xmm5 ; xmm1=tmp12 187 psubw xmm4,xmm5 ; xmm4=tmp3 189 paddw xmm6,xmm5 ; xmm6=tmp0 201 movdqa xmm5, XMMWORD [XMMBLOCK(5,0,esi,SIZEOF_JCOEF)] 203 pmullw xmm5, XMMWOR [all...] |
H A D | jfss2fst-64.asm | 110 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 112 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 123 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 128 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 130 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 136 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 137 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 140 movdqa xmm5, XMMWOR [all...] |
H A D | jfss2fst.asm | 116 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 118 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 129 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 134 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 136 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 142 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 143 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 146 movdqa xmm5, XMMWOR [all...] |
H A D | jiss2flt-64.asm | 106 movq xmm5, XMM_MMWORD [MMBLOCK(5,0,rsi,SIZEOF_JCOEF)] 111 por xmm5,xmm6 113 por xmm5,xmm7 114 por xmm1,xmm5 178 movaps xmm5,xmm1 182 addps xmm5,xmm3 ; xmm5=tmp13 185 subps xmm1,xmm5 ; xmm1=tmp12 189 subps xmm4,xmm5 ; xmm4=tmp3 191 addps xmm6,xmm5 ; xmm [all...] |
H A D | jfss2int-64.asm | 131 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 133 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 144 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 149 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 151 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 157 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 158 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 161 movdqa xmm5, XMMWOR [all...] |
H A D | jfss2int.asm | 137 movdqa xmm5,xmm2 ; transpose coefficients(phase 1) 139 punpckhwd xmm5,xmm3 ; xmm5=(24 34 25 35 26 36 27 37) 150 movdqa XMMWORD [wk(1)], xmm5 ; wk(1)=(24 34 25 35 26 36 27 37) 155 movdqa xmm5,xmm1 ; transpose coefficients(phase 1) 157 punpckhwd xmm5,xmm3 ; xmm5=(64 74 65 75 66 76 67 77) 163 punpckldq xmm2,xmm5 ; xmm2=(44 54 64 74 45 55 65 75) 164 punpckhdq xmm3,xmm5 ; xmm3=(46 56 66 76 47 57 67 77) 167 movdqa xmm5, XMMWOR [all...] |
H A D | jfsseflt-64.asm | 97 movaps xmm5,xmm2 ; transpose coefficients(phase 1) 99 unpckhps xmm5,xmm3 ; xmm5=(26 36 27 37) 123 unpcklps2 xmm2,xmm5 ; xmm2=(06 16 26 36)=data6 124 unpckhps2 xmm3,xmm5 ; xmm3=(07 17 27 37)=data7 127 movaps xmm5,xmm6 131 addps xmm5,xmm3 ; xmm5=data0+data7=tmp0 154 movaps xmm1,xmm5 156 subps xmm5,xmm [all...] |
H A D | jiss2int-64.asm | 147 movdqa xmm5, XMMWORD [XMMBLOCK(0,0,rsi,SIZEOF_JCOEF)] 148 pmullw xmm5, XMMWORD [XMMBLOCK(0,0,rdx,SIZEOF_ISLOW_MULT_TYPE)] 150 psllw xmm5,PASS1_BITS 152 movdqa xmm4,xmm5 ; xmm5=in0=(00 01 02 03 04 05 06 07) 153 punpcklwd xmm5,xmm5 ; xmm5=(00 00 01 01 02 02 03 03) 156 pshufd xmm7,xmm5,0x00 ; xmm7=col0=(00 00 00 00 00 00 00 00) 157 pshufd xmm6,xmm5, [all...] |
H A D | jiss2int.asm | 153 movdqa xmm5, XMMWORD [XMMBLOCK(0,0,esi,SIZEOF_JCOEF)] 154 pmullw xmm5, XMMWORD [XMMBLOCK(0,0,edx,SIZEOF_ISLOW_MULT_TYPE)] 156 psllw xmm5,PASS1_BITS 158 movdqa xmm4,xmm5 ; xmm5=in0=(00 01 02 03 04 05 06 07) 159 punpcklwd xmm5,xmm5 ; xmm5=(00 00 01 01 02 02 03 03) 162 pshufd xmm7,xmm5,0x00 ; xmm7=col0=(00 00 00 00 00 00 00 00) 163 pshufd xmm6,xmm5, [all...] |
/external/chromium_org/third_party/skia/src/opts/ |
H A D | SkBlitRow_opts_SSE4_asm.S | 30 movdqa %xmm5, %xmm3; \ 34 psllw $8, %xmm5; /* Filter out red and blue components */\ 35 pmulhuw %xmm4, %xmm5; /* Scale red and blue */\ 113 lddqu (%edx, %edi), %xmm5 // Load four destination pixels 119 pblendvb %xmm5, %xmm3 // Mask in %xmm0, implicitly 137 lddqu (%edx, %edi), %xmm5 // Load last four destination pixels (overlapping) 142 movdqa %xmm5, %xmm2 151 pblendw $0xF0, %xmm1, %xmm5 // Merge only the final two pixels to the destination 152 movdqu %xmm5, (%edx, %edi) // Store last two destination pixels 157 pblendw $0xC0, %xmm1, %xmm5 // Merg [all...] |
H A D | SkBlitRow_opts_SSE4_x64_asm.S | 18 movdqa %xmm5, %xmm3; \ 22 psllw $8, %xmm5; /* Filter out red and blue components */\ 23 pmulhuw %xmm4, %xmm5; /* Scale red and blue */\ 97 lddqu (%rdx, %rdi), %xmm5 // Load four destination pixels 103 pblendvb %xmm5, %xmm3 // Mask in %xmm0, implicitly 122 lddqu (%rdx, %rdi), %xmm5 // Load last four destination pixels (overlapping) 127 movdqa %xmm5, %xmm2 136 pblendw $0xF0, %xmm1, %xmm5 // Merge only the final two pixels to the destination 137 movdqu %xmm5, (%rdx, %rdi) // Store last two destination pixels 142 pblendw $0xC0, %xmm1, %xmm5 // Merg [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/x86/ |
H A D | vp9_sad_sse4.asm | 38 movq xmm5, MMWORD PTR [rdi] 41 punpcklqdq xmm5, xmm3 44 movdqa xmm2, xmm5 45 mpsadbw xmm5, xmm0, 0x0 54 paddw xmm5, xmm2 55 paddw xmm5, xmm3 56 paddw xmm5, xmm4 58 paddw xmm1, xmm5 61 movq xmm5, MMWORD PTR [rdi+ rdx] 64 punpcklqdq xmm5, xmm [all...] |