/external/llvm/test/MC/X86/ |
H A D | intel-syntax-invalid-scale.s | 7 lea rax, [rdi + rdx*64] 9 lea rax, [rdi + rdx*32] 11 lea rax, [rdi + rdx*16]
|
H A D | x86-64-avx512f_vl.s | 27 // CHECK: vaddpd 2032(%rdx), %xmm29, %xmm20 29 vaddpd 2032(%rdx), %xmm29, %xmm20 31 // CHECK: vaddpd 2048(%rdx), %xmm29, %xmm20 33 vaddpd 2048(%rdx), %xmm29, %xmm20 35 // CHECK: vaddpd -2048(%rdx), %xmm29, %xmm20 37 vaddpd -2048(%rdx), %xmm29, %xmm20 39 // CHECK: vaddpd -2064(%rdx), %xmm29, %xmm20 41 vaddpd -2064(%rdx), %xmm29, %xmm20 43 // CHECK: vaddpd 1016(%rdx){1to2}, %xmm29, %xmm20 45 vaddpd 1016(%rdx){ [all...] |
H A D | avx512vl-encoding.s | 27 // CHECK: vblendmpd 2032(%rdx), %xmm20, %xmm27 29 vblendmpd 2032(%rdx), %xmm20, %xmm27 31 // CHECK: vblendmpd 2048(%rdx), %xmm20, %xmm27 33 vblendmpd 2048(%rdx), %xmm20, %xmm27 35 // CHECK: vblendmpd -2048(%rdx), %xmm20, %xmm27 37 vblendmpd -2048(%rdx), %xmm20, %xmm27 39 // CHECK: vblendmpd -2064(%rdx), %xmm20, %xmm27 41 vblendmpd -2064(%rdx), %xmm20, %xmm27 43 // CHECK: vblendmpd 1016(%rdx){1to2}, %xmm20, %xmm27 45 vblendmpd 1016(%rdx){ [all...] |
H A D | x86-64-avx512bw_vl.s | 23 // CHECK: vpaddb 2032(%rdx), %xmm17, %xmm26 25 vpaddb 2032(%rdx), %xmm17, %xmm26 27 // CHECK: vpaddb 2048(%rdx), %xmm17, %xmm26 29 vpaddb 2048(%rdx), %xmm17, %xmm26 31 // CHECK: vpaddb -2048(%rdx), %xmm17, %xmm26 33 vpaddb -2048(%rdx), %xmm17, %xmm26 35 // CHECK: vpaddb -2064(%rdx), %xmm17, %xmm26 37 vpaddb -2064(%rdx), %xmm17, %xmm26 59 // CHECK: vpaddb 4064(%rdx), %ymm27, %ymm26 61 vpaddb 4064(%rdx), [all...] |
H A D | x86-64-avx512dq_vl.s | 27 // CHECK: vpmullq 2032(%rdx), %xmm17, %xmm26 29 vpmullq 2032(%rdx), %xmm17, %xmm26 31 // CHECK: vpmullq 2048(%rdx), %xmm17, %xmm26 33 vpmullq 2048(%rdx), %xmm17, %xmm26 35 // CHECK: vpmullq -2048(%rdx), %xmm17, %xmm26 37 vpmullq -2048(%rdx), %xmm17, %xmm26 39 // CHECK: vpmullq -2064(%rdx), %xmm17, %xmm26 41 vpmullq -2064(%rdx), %xmm17, %xmm26 43 // CHECK: vpmullq 1016(%rdx){1to2}, %xmm17, %xmm26 45 vpmullq 1016(%rdx){ [all...] |
H A D | avx512bw-encoding.s | 23 // CHECK: vpblendmb 8128(%rdx), %zmm18, %zmm17 25 vpblendmb 8128(%rdx), %zmm18, %zmm17 27 // CHECK: vpblendmb 8192(%rdx), %zmm18, %zmm17 29 vpblendmb 8192(%rdx), %zmm18, %zmm17 31 // CHECK: vpblendmb -8192(%rdx), %zmm18, %zmm17 33 vpblendmb -8192(%rdx), %zmm18, %zmm17 35 // CHECK: vpblendmb -8256(%rdx), %zmm18, %zmm17 37 vpblendmb -8256(%rdx), %zmm18, %zmm17 59 // CHECK: vpblendmw 8128(%rdx), %zmm20, %zmm26 61 vpblendmw 8128(%rdx), [all...] |
H A D | avx512-encodings.s | 28 // CHECK: vaddpd 8128(%rdx), %zmm27, %zmm8 30 vaddpd 8128(%rdx), %zmm27, %zmm8 32 // CHECK: vaddpd 8192(%rdx), %zmm27, %zmm8 34 vaddpd 8192(%rdx), %zmm27, %zmm8 36 // CHECK: vaddpd -8192(%rdx), %zmm27, %zmm8 38 vaddpd -8192(%rdx), %zmm27, %zmm8 40 // CHECK: vaddpd -8256(%rdx), %zmm27, %zmm8 42 vaddpd -8256(%rdx), %zmm27, %zmm8 44 // CHECK: vaddpd 1016(%rdx){1to8}, %zmm27, %zmm8 46 vaddpd 1016(%rdx){ [all...] |
H A D | x86-64-avx512bw.s | 23 // CHECK: vpaddb 8128(%rdx), %zmm24, %zmm19 25 vpaddb 8128(%rdx), %zmm24, %zmm19 27 // CHECK: vpaddb 8192(%rdx), %zmm24, %zmm19 29 vpaddb 8192(%rdx), %zmm24, %zmm19 31 // CHECK: vpaddb -8192(%rdx), %zmm24, %zmm19 33 vpaddb -8192(%rdx), %zmm24, %zmm19 35 // CHECK: vpaddb -8256(%rdx), %zmm24, %zmm19 37 vpaddb -8256(%rdx), %zmm24, %zmm19 59 // CHECK: vpaddw 8128(%rdx), %zmm24, %zmm17 61 vpaddw 8128(%rdx), [all...] |
H A D | x86-64-avx512dq.s | 27 // CHECK: vpmullq 8128(%rdx), %zmm24, %zmm18 29 vpmullq 8128(%rdx), %zmm24, %zmm18 31 // CHECK: vpmullq 8192(%rdx), %zmm24, %zmm18 33 vpmullq 8192(%rdx), %zmm24, %zmm18 35 // CHECK: vpmullq -8192(%rdx), %zmm24, %zmm18 37 vpmullq -8192(%rdx), %zmm24, %zmm18 39 // CHECK: vpmullq -8256(%rdx), %zmm24, %zmm18 41 vpmullq -8256(%rdx), %zmm24, %zmm18 43 // CHECK: vpmullq 1016(%rdx){1to8}, %zmm24, %zmm18 45 vpmullq 1016(%rdx){ [all...] |
/external/boringssl/linux-x86_64/crypto/bn/ |
H A D | modexp512-x86_64.S | 10 adcq $0,%rdx 12 movq %rdx,%rbx 17 adcq $0,%rdx 19 adcq $0,%rdx 20 movq %rdx,%rbx 25 adcq $0,%rdx 27 adcq $0,%rdx 28 movq %rdx,%rbx 33 adcq $0,%rdx 35 adcq $0,%rdx [all...] |
H A D | x86_64-mont.S | 16 cmpq %rsi,%rdx 40 movq %rdx,%r12 54 movq %rdx,%r11 59 adcq $0,%rdx 60 movq %rdx,%r13 69 adcq $0,%rdx 72 adcq $0,%rdx 74 movq %rdx,%r13 80 adcq $0,%rdx 82 movq %rdx, [all...] |
H A D | rsaz-x86_64.S | 21 movq %rdx,%rbp 22 movq (%rsi),%rdx 31 movq %rdx,%rbx 32 mulq %rdx 35 movq %rdx,%r9 40 movq %rdx,%r10 46 movq %rdx,%r11 52 movq %rdx,%r12 58 movq %rdx,%r13 64 movq %rdx, [all...] |
H A D | x86_64-mont5.S | 34 movq %rdx,%r12 85 movq %rdx,%r11 94 adcq $0,%rdx 95 movq %rdx,%r13 104 adcq $0,%rdx 107 adcq $0,%rdx 109 movq %rdx,%r13 115 adcq $0,%rdx 117 movq %rdx,%r10 127 adcq $0,%rdx [all...] |
/external/boringssl/mac-x86_64/crypto/bn/ |
H A D | modexp512-x86_64.S | 10 adcq $0,%rdx 12 movq %rdx,%rbx 17 adcq $0,%rdx 19 adcq $0,%rdx 20 movq %rdx,%rbx 25 adcq $0,%rdx 27 adcq $0,%rdx 28 movq %rdx,%rbx 33 adcq $0,%rdx 35 adcq $0,%rdx [all...] |
H A D | x86_64-mont.S | 15 cmpq %rsi,%rdx 39 movq %rdx,%r12 53 movq %rdx,%r11 58 adcq $0,%rdx 59 movq %rdx,%r13 68 adcq $0,%rdx 71 adcq $0,%rdx 73 movq %rdx,%r13 79 adcq $0,%rdx 81 movq %rdx, [all...] |
H A D | rsaz-x86_64.S | 20 movq %rdx,%rbp 21 movq (%rsi),%rdx 30 movq %rdx,%rbx 31 mulq %rdx 34 movq %rdx,%r9 39 movq %rdx,%r10 45 movq %rdx,%r11 51 movq %rdx,%r12 57 movq %rdx,%r13 63 movq %rdx, [all...] |
H A D | x86_64-mont5.S | 33 movq %rdx,%r12 84 movq %rdx,%r11 93 adcq $0,%rdx 94 movq %rdx,%r13 103 adcq $0,%rdx 106 adcq $0,%rdx 108 movq %rdx,%r13 114 adcq $0,%rdx 116 movq %rdx,%r10 126 adcq $0,%rdx [all...] |
/external/boringssl/win-x86_64/crypto/bn/ |
H A D | modexp512-x86_64.asm | 10 adc rdx,0 12 mov rbx,rdx 17 adc rdx,0 19 adc rdx,0 20 mov rbx,rdx 25 adc rdx,0 27 adc rdx,0 28 mov rbx,rdx 33 adc rdx,0 35 adc rdx, [all...] |
H A D | x86_64-mont.asm | 19 mov rsi,rdx 20 mov rdx,r8 30 cmp rdx,rsi 54 mov r12,rdx 68 mov r11,rdx 73 adc rdx,0 74 mov r13,rdx 83 adc rdx,0 86 adc rdx,0 88 mov r13,rdx [all...] |
/external/mesa3d/src/mesa/x86-64/ |
H A D | xform4.S | 55 * rdx = source 57 movl V4F_COUNT(%rdx), %ecx /* count */ 58 movzbl V4F_STRIDE(%rdx), %eax /* stride */ 69 movq V4F_START(%rdx), %rdx /* ptr to first src vertex */ 72 prefetch 16(%rdx) 82 movups (%rdx), %xmm8 /* ox | oy | oz | ow */ 86 addq %rax, %rdx 96 prefetch 16(%rdx) 140 movl V4F_COUNT(%rdx), [all...] |
/external/libunwind/src/x86_64/ |
H A D | longjmp.S | 30 mov %rdx, %rax /* set up return-value */
|
/external/valgrind/none/tests/amd64/ |
H A D | cmpxchg.c | 14 ULong rdx; variable 24 rdx = 0x11111111; rax = 0x22222222; 34 "\tpush %rdx\n" 40 "\tmov " VG_SYM(rdx) ",%rdx\n" 45 "\tmov " VG_SYM(rdx) "(%rip),%rdx\n" 57 "\tpop %rdx\n" 68 rdx = 0x99999999; rax = 0x77777777; 78 "\tpush %rdx\ [all...] |
/external/libvpx/libvpx/vp8/common/x86/ |
H A D | iwalsh_mmx.asm | 22 mov rdx, arg(0) 25 movq mm0, [rdx + 0] ;ip[0] 26 movq mm1, [rdx + 8] ;ip[4] 29 movq mm2, [rdx + 16] ;ip[8] 30 movq mm3, [rdx + 24] ;ip[12] 32 mov rdx, arg(1) 102 mov word ptr[rdx+32*0], ax 103 mov word ptr[rdx+32*1], cx 106 mov word ptr[rdx+32*4], ax 107 mov word ptr[rdx [all...] |
/external/v8/test/cctest/ |
H A D | test-disasm-x64.cc | 69 __ addq(rdx, rbx); 70 __ addq(rdx, Operand(rbx, 0)); 71 __ addq(rdx, Operand(rbx, 16)); 72 __ addq(rdx, Operand(rbx, 1999)); 73 __ addq(rdx, Operand(rbx, -4)); 74 __ addq(rdx, Operand(rbx, -1999)); 75 __ addq(rdx, Operand(rsp, 0)); 76 __ addq(rdx, Operand(rsp, 16)); 77 __ addq(rdx, Operand(rsp, 1999)); 78 __ addq(rdx, Operan 140 __ movb(Operand(rsp, 16), rdx); local 141 __ movw(Operand(rsp, 16), rdx); local 152 __ movq(Operand(rbx, rcx, times_4, 10000), rdx); local 165 __ testq(Operand(rbx, rcx, times_4, 10000), rdx); local 230 __ testq(Operand(rbx, rcx, times_8, 10000), rdx); local 231 __ testb(Operand(rcx, rbx, times_2, 1000), rdx); local 237 __ bts(Operand(rbx, rcx, times_8, 10000), rdx); local [all...] |
/external/zlib/src/contrib/amd64/ |
H A D | amd64-match.S | 269 mov $(-MAX_MATCH_8), %rdx 275 prefetcht1 (%windowbestlen, %rdx) 276 prefetcht1 (%prev, %rdx) 280 * adjust %rdx so that it is offset to the exact byte that mismatched. 293 movdqu (%windowbestlen, %rdx), %xmm1 294 movdqu (%prev, %rdx), %xmm2 296 movdqu 16(%windowbestlen, %rdx), %xmm3 297 movdqu 16(%prev, %rdx), %xmm4 299 movdqu 32(%windowbestlen, %rdx), %xmm5 300 movdqu 32(%prev, %rdx), [all...] |