Searched refs:rdx (Results 1 - 25 of 220) sorted by relevance

123456789

/external/llvm/test/MC/X86/
H A Dintel-syntax-invalid-scale.s7 lea rax, [rdi + rdx*64]
9 lea rax, [rdi + rdx*32]
11 lea rax, [rdi + rdx*16]
H A Dx86-64-avx512cd.s27 // CHECK: vplzcntq 4064(%rdx), %zmm21
29 vplzcntq 4064(%rdx), %zmm21
31 // CHECK: vplzcntq 4096(%rdx), %zmm21
33 vplzcntq 4096(%rdx), %zmm21
35 // CHECK: vplzcntq -4096(%rdx), %zmm21
37 vplzcntq -4096(%rdx), %zmm21
39 // CHECK: vplzcntq -4128(%rdx), %zmm21
41 vplzcntq -4128(%rdx), %zmm21
43 // CHECK: vplzcntq 1016(%rdx){1to8}, %zmm21
45 vplzcntq 1016(%rdx){
[all...]
H A Dx86-64-avx512cd_vl.s27 // CHECK: vplzcntq 2032(%rdx), %xmm18
29 vplzcntq 2032(%rdx), %xmm18
31 // CHECK: vplzcntq 2048(%rdx), %xmm18
33 vplzcntq 2048(%rdx), %xmm18
35 // CHECK: vplzcntq -2048(%rdx), %xmm18
37 vplzcntq -2048(%rdx), %xmm18
39 // CHECK: vplzcntq -2064(%rdx), %xmm18
41 vplzcntq -2064(%rdx), %xmm18
43 // CHECK: vplzcntq 1016(%rdx){1to2}, %xmm18
45 vplzcntq 1016(%rdx){
[all...]
H A Dx86-64-avx512dq_vl.s27 // CHECK: vpmullq 2032(%rdx), %xmm17, %xmm26
29 vpmullq 2032(%rdx), %xmm17, %xmm26
31 // CHECK: vpmullq 2048(%rdx), %xmm17, %xmm26
33 vpmullq 2048(%rdx), %xmm17, %xmm26
35 // CHECK: vpmullq -2048(%rdx), %xmm17, %xmm26
37 vpmullq -2048(%rdx), %xmm17, %xmm26
39 // CHECK: vpmullq -2064(%rdx), %xmm17, %xmm26
41 vpmullq -2064(%rdx), %xmm17, %xmm26
43 // CHECK: vpmullq 1016(%rdx){1to2}, %xmm17, %xmm26
45 vpmullq 1016(%rdx){
[all...]
H A Davx512bw-encoding.s23 // CHECK: vpblendmb 8128(%rdx), %zmm18, %zmm17
25 vpblendmb 8128(%rdx), %zmm18, %zmm17
27 // CHECK: vpblendmb 8192(%rdx), %zmm18, %zmm17
29 vpblendmb 8192(%rdx), %zmm18, %zmm17
31 // CHECK: vpblendmb -8192(%rdx), %zmm18, %zmm17
33 vpblendmb -8192(%rdx), %zmm18, %zmm17
35 // CHECK: vpblendmb -8256(%rdx), %zmm18, %zmm17
37 vpblendmb -8256(%rdx), %zmm18, %zmm17
59 // CHECK: vpblendmw 8128(%rdx), %zmm20, %zmm26
61 vpblendmw 8128(%rdx),
[all...]
H A Dx86-64-avx512bw.s23 // CHECK: vpaddb 8128(%rdx), %zmm24, %zmm19
25 vpaddb 8128(%rdx), %zmm24, %zmm19
27 // CHECK: vpaddb 8192(%rdx), %zmm24, %zmm19
29 vpaddb 8192(%rdx), %zmm24, %zmm19
31 // CHECK: vpaddb -8192(%rdx), %zmm24, %zmm19
33 vpaddb -8192(%rdx), %zmm24, %zmm19
35 // CHECK: vpaddb -8256(%rdx), %zmm24, %zmm19
37 vpaddb -8256(%rdx), %zmm24, %zmm19
59 // CHECK: vpaddw 8128(%rdx), %zmm24, %zmm17
61 vpaddw 8128(%rdx),
[all...]
H A Dx86-64-avx512f_vl.s27 // CHECK: vaddpd 2032(%rdx), %xmm29, %xmm20
29 vaddpd 2032(%rdx), %xmm29, %xmm20
31 // CHECK: vaddpd 2048(%rdx), %xmm29, %xmm20
33 vaddpd 2048(%rdx), %xmm29, %xmm20
35 // CHECK: vaddpd -2048(%rdx), %xmm29, %xmm20
37 vaddpd -2048(%rdx), %xmm29, %xmm20
39 // CHECK: vaddpd -2064(%rdx), %xmm29, %xmm20
41 vaddpd -2064(%rdx), %xmm29, %xmm20
43 // CHECK: vaddpd 1016(%rdx){1to2}, %xmm29, %xmm20
45 vaddpd 1016(%rdx){
[all...]
H A Dx86-64-avx512dq.s27 // CHECK: vpmullq 8128(%rdx), %zmm24, %zmm18
29 vpmullq 8128(%rdx), %zmm24, %zmm18
31 // CHECK: vpmullq 8192(%rdx), %zmm24, %zmm18
33 vpmullq 8192(%rdx), %zmm24, %zmm18
35 // CHECK: vpmullq -8192(%rdx), %zmm24, %zmm18
37 vpmullq -8192(%rdx), %zmm24, %zmm18
39 // CHECK: vpmullq -8256(%rdx), %zmm24, %zmm18
41 vpmullq -8256(%rdx), %zmm24, %zmm18
43 // CHECK: vpmullq 1016(%rdx){1to8}, %zmm24, %zmm18
45 vpmullq 1016(%rdx){
[all...]
H A Davx512vl-encoding.s27 // CHECK: vblendmpd 2032(%rdx), %xmm20, %xmm27
29 vblendmpd 2032(%rdx), %xmm20, %xmm27
31 // CHECK: vblendmpd 2048(%rdx), %xmm20, %xmm27
33 vblendmpd 2048(%rdx), %xmm20, %xmm27
35 // CHECK: vblendmpd -2048(%rdx), %xmm20, %xmm27
37 vblendmpd -2048(%rdx), %xmm20, %xmm27
39 // CHECK: vblendmpd -2064(%rdx), %xmm20, %xmm27
41 vblendmpd -2064(%rdx), %xmm20, %xmm27
43 // CHECK: vblendmpd 1016(%rdx){1to2}, %xmm20, %xmm27
45 vblendmpd 1016(%rdx){
[all...]
H A Dintel-syntax-x86-64-avx512f_vl.s27 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx + 2032], 123
29 vcmppd k3,xmm27,XMMWORD PTR [rdx+0x7f0],0x7b
31 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx + 2048], 123
33 vcmppd k3,xmm27,XMMWORD PTR [rdx+0x800],0x7b
35 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx - 2048], 123
37 vcmppd k3,xmm27,XMMWORD PTR [rdx-0x800],0x7b
39 // CHECK: vcmppd k3 , xmm27, xmmword ptr [rdx - 2064], 123
41 vcmppd k3,xmm27,XMMWORD PTR [rdx-0x810],0x7b
43 // CHECK: vcmppd k3 , xmm27, qword ptr [rdx + 1016]{1to2}, 123
45 vcmppd k3,xmm27,QWORD PTR [rdx
[all...]
H A Dx86-64-avx512bw_vl.s23 // CHECK: vpaddb 2032(%rdx), %xmm17, %xmm26
25 vpaddb 2032(%rdx), %xmm17, %xmm26
27 // CHECK: vpaddb 2048(%rdx), %xmm17, %xmm26
29 vpaddb 2048(%rdx), %xmm17, %xmm26
31 // CHECK: vpaddb -2048(%rdx), %xmm17, %xmm26
33 vpaddb -2048(%rdx), %xmm17, %xmm26
35 // CHECK: vpaddb -2064(%rdx), %xmm17, %xmm26
37 vpaddb -2064(%rdx), %xmm17, %xmm26
59 // CHECK: vpaddb 4064(%rdx), %ymm27, %ymm26
61 vpaddb 4064(%rdx),
[all...]
H A Davx512-encodings.s27 // CHECK: vaddpd 8128(%rdx), %zmm27, %zmm8
29 vaddpd 8128(%rdx), %zmm27, %zmm8
31 // CHECK: vaddpd 8192(%rdx), %zmm27, %zmm8
33 vaddpd 8192(%rdx), %zmm27, %zmm8
35 // CHECK: vaddpd -8192(%rdx), %zmm27, %zmm8
37 vaddpd -8192(%rdx), %zmm27, %zmm8
39 // CHECK: vaddpd -8256(%rdx), %zmm27, %zmm8
41 vaddpd -8256(%rdx), %zmm27, %zmm8
43 // CHECK: vaddpd 1016(%rdx){1to8}, %zmm27, %zmm8
45 vaddpd 1016(%rdx){
[all...]
H A Dintel-syntax-avx512.s67 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx + 8128], 123
69 vcmppd k2,zmm12,zmmword PTR [rdx+0x1fc0],0x7b
71 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx + 8192], 123
73 vcmppd k2,zmm12,zmmword PTR [rdx+0x2000],0x7b
75 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx - 8192], 123
77 vcmppd k2,zmm12,zmmword PTR [rdx-0x2000],0x7b
79 // CHECK: vcmppd k2 , zmm12, zmmword ptr [rdx - 8256], 123
81 vcmppd k2,zmm12,zmmword PTR [rdx-0x2040],0x7b
83 // CHECK: vcmppd k2 , zmm12, qword ptr [rdx + 1016]{1to8}, 123
85 vcmppd k2,zmm12,QWORD PTR [rdx
[all...]
/external/boringssl/linux-x86_64/crypto/bn/
H A Dx86_64-mont.S16 cmpq %rsi,%rdx
40 movq %rdx,%r12
54 movq %rdx,%r11
59 adcq $0,%rdx
60 movq %rdx,%r13
69 adcq $0,%rdx
72 adcq $0,%rdx
74 movq %rdx,%r13
80 adcq $0,%rdx
82 movq %rdx,
[all...]
H A Drsaz-x86_64.S21 movq %rdx,%rbp
22 movq (%rsi),%rdx
31 movq %rdx,%rbx
32 mulq %rdx
35 movq %rdx,%r9
40 movq %rdx,%r10
46 movq %rdx,%r11
52 movq %rdx,%r12
58 movq %rdx,%r13
64 movq %rdx,
[all...]
H A Dx86_64-mont5.S34 movq %rdx,%r12
85 movq %rdx,%r11
94 adcq $0,%rdx
95 movq %rdx,%r13
104 adcq $0,%rdx
107 adcq $0,%rdx
109 movq %rdx,%r13
115 adcq $0,%rdx
117 movq %rdx,%r10
127 adcq $0,%rdx
[all...]
/external/boringssl/mac-x86_64/crypto/bn/
H A Dx86_64-mont.S15 cmpq %rsi,%rdx
39 movq %rdx,%r12
53 movq %rdx,%r11
58 adcq $0,%rdx
59 movq %rdx,%r13
68 adcq $0,%rdx
71 adcq $0,%rdx
73 movq %rdx,%r13
79 adcq $0,%rdx
81 movq %rdx,
[all...]
H A Drsaz-x86_64.S20 movq %rdx,%rbp
21 movq (%rsi),%rdx
30 movq %rdx,%rbx
31 mulq %rdx
34 movq %rdx,%r9
39 movq %rdx,%r10
45 movq %rdx,%r11
51 movq %rdx,%r12
57 movq %rdx,%r13
63 movq %rdx,
[all...]
/external/boringssl/src/crypto/curve25519/asm/
H A Dx25519-asm-x86_64.S72 movq 8(%rdi),%rdx
84 add %r12,%rdx
85 mov %rdx,%r12
87 and %rax,%rdx
107 cmp %rax,%rdx
119 sub %rax,%rdx
124 movq %rdx,8(%rdi)
137 mov %rsi,%rdx
156 mov %rdx,%rcx
157 movq 24(%rsi),%rdx
[all...]
/external/mesa3d/src/mesa/x86-64/
H A Dxform4.S55 * rdx = source
57 movl V4F_COUNT(%rdx), %ecx /* count */
58 movzbl V4F_STRIDE(%rdx), %eax /* stride */
69 movq V4F_START(%rdx), %rdx /* ptr to first src vertex */
72 prefetch 16(%rdx)
82 movups (%rdx), %xmm8 /* ox | oy | oz | ow */
86 addq %rax, %rdx
96 prefetch 16(%rdx)
140 movl V4F_COUNT(%rdx),
[all...]
/external/boringssl/win-x86_64/crypto/bn/
H A Dx86_64-mont.asm19 mov rsi,rdx
20 mov rdx,r8
30 cmp rdx,rsi
54 mov r12,rdx
68 mov r11,rdx
73 adc rdx,0
74 mov r13,rdx
83 adc rdx,0
86 adc rdx,0
88 mov r13,rdx
[all...]
/external/libunwind/src/x86_64/
H A Dlongjmp.S30 mov %rdx, %rax /* set up return-value */
/external/valgrind/none/tests/amd64/
H A Dcmpxchg.c14 ULong rdx; variable
24 rdx = 0x11111111; rax = 0x22222222;
34 "\tpush %rdx\n"
40 "\tmov " VG_SYM(rdx) ",%rdx\n"
45 "\tmov " VG_SYM(rdx) "(%rip),%rdx\n"
57 "\tpop %rdx\n"
68 rdx = 0x99999999; rax = 0x77777777;
78 "\tpush %rdx\
[all...]
/external/libvpx/libvpx/vp8/common/x86/
H A Diwalsh_mmx.asm22 mov rdx, arg(0)
25 movq mm0, [rdx + 0] ;ip[0]
26 movq mm1, [rdx + 8] ;ip[4]
29 movq mm2, [rdx + 16] ;ip[8]
30 movq mm3, [rdx + 24] ;ip[12]
32 mov rdx, arg(1)
102 mov word ptr[rdx+32*0], ax
103 mov word ptr[rdx+32*1], cx
106 mov word ptr[rdx+32*4], ax
107 mov word ptr[rdx
[all...]
/external/zlib/src/contrib/amd64/
H A Damd64-match.S269 mov $(-MAX_MATCH_8), %rdx
275 prefetcht1 (%windowbestlen, %rdx)
276 prefetcht1 (%prev, %rdx)
280 * adjust %rdx so that it is offset to the exact byte that mismatched.
293 movdqu (%windowbestlen, %rdx), %xmm1
294 movdqu (%prev, %rdx), %xmm2
296 movdqu 16(%windowbestlen, %rdx), %xmm3
297 movdqu 16(%prev, %rdx), %xmm4
299 movdqu 32(%windowbestlen, %rdx), %xmm5
300 movdqu 32(%prev, %rdx),
[all...]

Completed in 796 milliseconds

123456789