/external/llvm/test/MC/X86/ |
H A D | intel-syntax-invalid-scale.s | 7 lea rax, [rdi + rdx*64] 9 lea rax, [rdi + rdx*32] 11 lea rax, [rdi + rdx*16]
|
H A D | x86-64-avx512cd.s | 27 // CHECK: vplzcntq 4064(%rdx), %zmm21 29 vplzcntq 4064(%rdx), %zmm21 31 // CHECK: vplzcntq 4096(%rdx), %zmm21 33 vplzcntq 4096(%rdx), %zmm21 35 // CHECK: vplzcntq -4096(%rdx), %zmm21 37 vplzcntq -4096(%rdx), %zmm21 39 // CHECK: vplzcntq -4128(%rdx), %zmm21 41 vplzcntq -4128(%rdx), %zmm21 43 // CHECK: vplzcntq 1016(%rdx){1to8}, %zmm21 45 vplzcntq 1016(%rdx){ [all...] |
H A D | x86-64-avx512cd_vl.s | 27 // CHECK: vplzcntq 2032(%rdx), %xmm18 29 vplzcntq 2032(%rdx), %xmm18 31 // CHECK: vplzcntq 2048(%rdx), %xmm18 33 vplzcntq 2048(%rdx), %xmm18 35 // CHECK: vplzcntq -2048(%rdx), %xmm18 37 vplzcntq -2048(%rdx), %xmm18 39 // CHECK: vplzcntq -2064(%rdx), %xmm18 41 vplzcntq -2064(%rdx), %xmm18 43 // CHECK: vplzcntq 1016(%rdx){1to2}, %xmm18 45 vplzcntq 1016(%rdx){ [all...] |
H A D | x86-64-avx512dq_vl.s | 27 // CHECK: vpmullq 2032(%rdx), %xmm17, %xmm26 29 vpmullq 2032(%rdx), %xmm17, %xmm26 31 // CHECK: vpmullq 2048(%rdx), %xmm17, %xmm26 33 vpmullq 2048(%rdx), %xmm17, %xmm26 35 // CHECK: vpmullq -2048(%rdx), %xmm17, %xmm26 37 vpmullq -2048(%rdx), %xmm17, %xmm26 39 // CHECK: vpmullq -2064(%rdx), %xmm17, %xmm26 41 vpmullq -2064(%rdx), %xmm17, %xmm26 43 // CHECK: vpmullq 1016(%rdx){1to2}, %xmm17, %xmm26 45 vpmullq 1016(%rdx){ [all...] |
H A D | avx512bw-encoding.s | 23 // CHECK: vpblendmb 8128(%rdx), %zmm18, %zmm17 25 vpblendmb 8128(%rdx), %zmm18, %zmm17 27 // CHECK: vpblendmb 8192(%rdx), %zmm18, %zmm17 29 vpblendmb 8192(%rdx), %zmm18, %zmm17 31 // CHECK: vpblendmb -8192(%rdx), %zmm18, %zmm17 33 vpblendmb -8192(%rdx), %zmm18, %zmm17 35 // CHECK: vpblendmb -8256(%rdx), %zmm18, %zmm17 37 vpblendmb -8256(%rdx), %zmm18, %zmm17 59 // CHECK: vpblendmw 8128(%rdx), %zmm20, %zmm26 61 vpblendmw 8128(%rdx), [all...] |
H A D | x86-64-avx512bw.s | 23 // CHECK: vpaddb 8128(%rdx), %zmm24, %zmm19 25 vpaddb 8128(%rdx), %zmm24, %zmm19 27 // CHECK: vpaddb 8192(%rdx), %zmm24, %zmm19 29 vpaddb 8192(%rdx), %zmm24, %zmm19 31 // CHECK: vpaddb -8192(%rdx), %zmm24, %zmm19 33 vpaddb -8192(%rdx), %zmm24, %zmm19 35 // CHECK: vpaddb -8256(%rdx), %zmm24, %zmm19 37 vpaddb -8256(%rdx), %zmm24, %zmm19 59 // CHECK: vpaddw 8128(%rdx), %zmm24, %zmm17 61 vpaddw 8128(%rdx), [all...] |
H A D | x86-64-avx512f_vl.s | 27 // CHECK: vaddpd 2032(%rdx), %xmm29, %xmm20 29 vaddpd 2032(%rdx), %xmm29, %xmm20 31 // CHECK: vaddpd 2048(%rdx), %xmm29, %xmm20 33 vaddpd 2048(%rdx), %xmm29, %xmm20 35 // CHECK: vaddpd -2048(%rdx), %xmm29, %xmm20 37 vaddpd -2048(%rdx), %xmm29, %xmm20 39 // CHECK: vaddpd -2064(%rdx), %xmm29, %xmm20 41 vaddpd -2064(%rdx), %xmm29, %xmm20 43 // CHECK: vaddpd 1016(%rdx){1to2}, %xmm29, %xmm20 45 vaddpd 1016(%rdx){ [all...] |
H A D | x86-64-avx512dq.s | 27 // CHECK: vpmullq 8128(%rdx), %zmm24, %zmm18 29 vpmullq 8128(%rdx), %zmm24, %zmm18 31 // CHECK: vpmullq 8192(%rdx), %zmm24, %zmm18 33 vpmullq 8192(%rdx), %zmm24, %zmm18 35 // CHECK: vpmullq -8192(%rdx), %zmm24, %zmm18 37 vpmullq -8192(%rdx), %zmm24, %zmm18 39 // CHECK: vpmullq -8256(%rdx), %zmm24, %zmm18 41 vpmullq -8256(%rdx), %zmm24, %zmm18 43 // CHECK: vpmullq 1016(%rdx){1to8}, %zmm24, %zmm18 45 vpmullq 1016(%rdx){ [all...] |
H A D | avx512vl-encoding.s | 27 // CHECK: vblendmpd 2032(%rdx), %xmm20, %xmm27 29 vblendmpd 2032(%rdx), %xmm20, %xmm27 31 // CHECK: vblendmpd 2048(%rdx), %xmm20, %xmm27 33 vblendmpd 2048(%rdx), %xmm20, %xmm27 35 // CHECK: vblendmpd -2048(%rdx), %xmm20, %xmm27 37 vblendmpd -2048(%rdx), %xmm20, %xmm27 39 // CHECK: vblendmpd -2064(%rdx), %xmm20, %xmm27 41 vblendmpd -2064(%rdx), %xmm20, %xmm27 43 // CHECK: vblendmpd 1016(%rdx){1to2}, %xmm20, %xmm27 45 vblendmpd 1016(%rdx){ [all...] |
H A D | intel-syntax-x86-64-avx512f_vl.s | 27 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx + 2032], 123 29 vcmppd k3,xmm27,XMMWORD PTR [rdx+0x7f0],0x7b 31 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx + 2048], 123 33 vcmppd k3,xmm27,XMMWORD PTR [rdx+0x800],0x7b 35 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx - 2048], 123 37 vcmppd k3,xmm27,XMMWORD PTR [rdx-0x800],0x7b 39 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx - 2064], 123 41 vcmppd k3,xmm27,XMMWORD PTR [rdx-0x810],0x7b 43 // CHECK: vcmppd k3 , xmm27, qword ptr [rdx + 1016]{1to2}, 123 45 vcmppd k3,xmm27,QWORD PTR [rdx [all...] |
H A D | x86-64-avx512bw_vl.s | 23 // CHECK: vpaddb 2032(%rdx), %xmm17, %xmm26 25 vpaddb 2032(%rdx), %xmm17, %xmm26 27 // CHECK: vpaddb 2048(%rdx), %xmm17, %xmm26 29 vpaddb 2048(%rdx), %xmm17, %xmm26 31 // CHECK: vpaddb -2048(%rdx), %xmm17, %xmm26 33 vpaddb -2048(%rdx), %xmm17, %xmm26 35 // CHECK: vpaddb -2064(%rdx), %xmm17, %xmm26 37 vpaddb -2064(%rdx), %xmm17, %xmm26 59 // CHECK: vpaddb 4064(%rdx), %ymm27, %ymm26 61 vpaddb 4064(%rdx), [all...] |
H A D | avx512-encodings.s | 27 // CHECK: vaddpd 8128(%rdx), %zmm27, %zmm8 29 vaddpd 8128(%rdx), %zmm27, %zmm8 31 // CHECK: vaddpd 8192(%rdx), %zmm27, %zmm8 33 vaddpd 8192(%rdx), %zmm27, %zmm8 35 // CHECK: vaddpd -8192(%rdx), %zmm27, %zmm8 37 vaddpd -8192(%rdx), %zmm27, %zmm8 39 // CHECK: vaddpd -8256(%rdx), %zmm27, %zmm8 41 vaddpd -8256(%rdx), %zmm27, %zmm8 43 // CHECK: vaddpd 1016(%rdx){1to8}, %zmm27, %zmm8 45 vaddpd 1016(%rdx){ [all...] |
H A D | intel-syntax-avx512.s | 67 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx + 8128], 123 69 vcmppd k2,zmm12,zmmword PTR [rdx+0x1fc0],0x7b 71 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx + 8192], 123 73 vcmppd k2,zmm12,zmmword PTR [rdx+0x2000],0x7b 75 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx - 8192], 123 77 vcmppd k2,zmm12,zmmword PTR [rdx-0x2000],0x7b 79 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx - 8256], 123 81 vcmppd k2,zmm12,zmmword PTR [rdx-0x2040],0x7b 83 // CHECK: vcmppd k2 , zmm12, qword ptr [rdx + 1016]{1to8}, 123 85 vcmppd k2,zmm12,QWORD PTR [rdx [all...] |
/external/boringssl/linux-x86_64/crypto/bn/ |
H A D | x86_64-mont.S | 16 cmpq %rsi,%rdx 40 movq %rdx,%r12 54 movq %rdx,%r11 59 adcq $0,%rdx 60 movq %rdx,%r13 69 adcq $0,%rdx 72 adcq $0,%rdx 74 movq %rdx,%r13 80 adcq $0,%rdx 82 movq %rdx, [all...] |
H A D | rsaz-x86_64.S | 21 movq %rdx,%rbp 22 movq (%rsi),%rdx 31 movq %rdx,%rbx 32 mulq %rdx 35 movq %rdx,%r9 40 movq %rdx,%r10 46 movq %rdx,%r11 52 movq %rdx,%r12 58 movq %rdx,%r13 64 movq %rdx, [all...] |
H A D | x86_64-mont5.S | 34 movq %rdx,%r12 85 movq %rdx,%r11 94 adcq $0,%rdx 95 movq %rdx,%r13 104 adcq $0,%rdx 107 adcq $0,%rdx 109 movq %rdx,%r13 115 adcq $0,%rdx 117 movq %rdx,%r10 127 adcq $0,%rdx [all...] |
/external/boringssl/mac-x86_64/crypto/bn/ |
H A D | x86_64-mont.S | 15 cmpq %rsi,%rdx 39 movq %rdx,%r12 53 movq %rdx,%r11 58 adcq $0,%rdx 59 movq %rdx,%r13 68 adcq $0,%rdx 71 adcq $0,%rdx 73 movq %rdx,%r13 79 adcq $0,%rdx 81 movq %rdx, [all...] |
H A D | rsaz-x86_64.S | 20 movq %rdx,%rbp 21 movq (%rsi),%rdx 30 movq %rdx,%rbx 31 mulq %rdx 34 movq %rdx,%r9 39 movq %rdx,%r10 45 movq %rdx,%r11 51 movq %rdx,%r12 57 movq %rdx,%r13 63 movq %rdx, [all...] |
/external/boringssl/src/crypto/curve25519/asm/ |
H A D | x25519-asm-x86_64.S | 72 movq 8(%rdi),%rdx 84 add %r12,%rdx 85 mov %rdx,%r12 87 and %rax,%rdx 107 cmp %rax,%rdx 119 sub %rax,%rdx 124 movq %rdx,8(%rdi) 137 mov %rsi,%rdx 156 mov %rdx,%rcx 157 movq 24(%rsi),%rdx [all...] |
/external/mesa3d/src/mesa/x86-64/ |
H A D | xform4.S | 55 * rdx = source 57 movl V4F_COUNT(%rdx), %ecx /* count */ 58 movzbl V4F_STRIDE(%rdx), %eax /* stride */ 69 movq V4F_START(%rdx), %rdx /* ptr to first src vertex */ 72 prefetch 16(%rdx) 82 movups (%rdx), %xmm8 /* ox | oy | oz | ow */ 86 addq %rax, %rdx 96 prefetch 16(%rdx) 140 movl V4F_COUNT(%rdx), [all...] |
/external/boringssl/win-x86_64/crypto/bn/ |
H A D | x86_64-mont.asm | 19 mov rsi,rdx 20 mov rdx,r8 30 cmp rdx,rsi 54 mov r12,rdx 68 mov r11,rdx 73 adc rdx,0 74 mov r13,rdx 83 adc rdx,0 86 adc rdx,0 88 mov r13,rdx [all...] |
/external/libunwind/src/x86_64/ |
H A D | longjmp.S | 30 mov %rdx, %rax /* set up return-value */
|
/external/valgrind/none/tests/amd64/ |
H A D | cmpxchg.c | 14 ULong rdx; variable 24 rdx = 0x11111111; rax = 0x22222222; 34 "\tpush %rdx\n" 40 "\tmov " VG_SYM(rdx) ",%rdx\n" 45 "\tmov " VG_SYM(rdx) "(%rip),%rdx\n" 57 "\tpop %rdx\n" 68 rdx = 0x99999999; rax = 0x77777777; 78 "\tpush %rdx\ [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
H A D | iwalsh_mmx.asm | 22 mov rdx, arg(0) 25 movq mm0, [rdx + 0] ;ip[0] 26 movq mm1, [rdx + 8] ;ip[4] 29 movq mm2, [rdx + 16] ;ip[8] 30 movq mm3, [rdx + 24] ;ip[12] 32 mov rdx, arg(1) 102 mov word ptr[rdx+32*0], ax 103 mov word ptr[rdx+32*1], cx 106 mov word ptr[rdx+32*4], ax 107 mov word ptr[rdx [all...] |
/external/zlib/src/contrib/amd64/ |
H A D | amd64-match.S | 269 mov $(-MAX_MATCH_8), %rdx 275 prefetcht1 (%windowbestlen, %rdx) 276 prefetcht1 (%prev, %rdx) 280 * adjust %rdx so that it is offset to the exact byte that mismatched. 293 movdqu (%windowbestlen, %rdx), %xmm1 294 movdqu (%prev, %rdx), %xmm2 296 movdqu 16(%windowbestlen, %rdx), %xmm3 297 movdqu 16(%prev, %rdx), %xmm4 299 movdqu 32(%windowbestlen, %rdx), %xmm5 300 movdqu 32(%prev, %rdx), [all...] |