/external/llvm/test/MC/ARM/ |
H A D | ldrd-strd-gnu-arm.s | 8 @ CHECK: ldrd r0, r1, [r10, #32]! @ encoding: [0xd0,0x02,0xea,0xe1] 9 @ CHECK: ldrd r0, r1, [r10], #32 @ encoding: [0xd0,0x02,0xca,0xe0] 10 @ CHECK: ldrd r0, r1, [r10, #32] @ encoding: [0xd0,0x02,0xca,0xe1] 11 ldrd r0, [r10, #32]! 12 ldrd r0, [r10], #32 13 ldrd r0, [r10, #32] 15 @ CHECK: strd r0, r1, [r10, #32]! @ encoding: [0xf0,0x02,0xea,0xe1] 16 @ CHECK: strd r0, r1, [r10], #32 @ encoding: [0xf0,0x02,0xca,0xe0] 17 @ CHECK: strd r0, r1, [r10, #32] @ encoding: [0xf0,0x02,0xca,0xe1] 18 strd r0, [r10, #3 [all...] |
H A D | ldrd-strd-gnu-thumb.s | 8 @ CHECK: ldrd r0, r1, [r10, #512]! @ encoding: [0xfa,0xe9,0x80,0x01] 9 @ CHECK: ldrd r0, r1, [r10], #512 @ encoding: [0xfa,0xe8,0x80,0x01] 10 @ CHECK: ldrd r0, r1, [r10, #512] @ encoding: [0xda,0xe9,0x80,0x01] 11 ldrd r0, [r10, #512]! 12 ldrd r0, [r10], #512 13 ldrd r0, [r10, #512] 15 @ CHECK: strd r0, r1, [r10, #512]! @ encoding: [0xea,0xe9,0x80,0x01] 16 @ CHECK: strd r0, r1, [r10], #512 @ encoding: [0xea,0xe8,0x80,0x01] 17 @ CHECK: strd r0, r1, [r10, #512] @ encoding: [0xca,0xe9,0x80,0x01] 18 strd r0, [r10, #51 [all...] |
/external/llvm/test/MC/X86/ |
H A D | x86_64-bmi-encoding.s | 7 // CHECK: blsmskq %r11, %r10 9 blsmskq %r11, %r10 15 // CHECK: blsmskq (%rax), %r10 17 blsmskq (%rax), %r10 23 // CHECK: blsiq %r11, %r10 25 blsiq %r11, %r10 31 // CHECK: blsiq (%rax), %r10 33 blsiq (%rax), %r10 39 // CHECK: blsrq %r11, %r10 41 blsrq %r11, %r10 [all...] |
/external/valgrind/coregrind/m_mach/ |
H A D | mach_traps-amd64-darwin.S | 42 movq %rcx, %r10 53 // movq %rcx, %r10 63 movq %rcx, %r10 73 movq %rcx, %r10 83 movq %rcx, %r10 93 movq %rcx, %r10 103 movq %rcx, %r10 113 movq %rcx, %r10 123 movq %rcx, %r10 133 movq %rcx, %r10 [all...] |
/external/python/cpython2/Modules/_ctypes/libffi/src/tile/ |
H A D | tile.S | 83 #define FRAME_SIZE r10 231 On entry, lr points to the closure plus 8 bytes, and r10 252 /* Save return address (in r10 due to closure stub wrapper). */ 253 SW sp, r10 254 .cfi_return_column r10 255 .cfi_offset r10, 0 258 addli r10, sp, -(CLOSURE_FRAME_SIZE - REG_SIZE) 262 SW r10, sp 271 addi r10, sp, LINKAGE_SIZE 275 STORE_REG(r0, r10) [all...] |
/external/libhevc/common/arm/ |
H A D | ihevc_intra_pred_luma_mode_18_34.s | 126 mov r10,r2 160 vst1.8 {d0},[r10],r3 161 vst1.8 {d1},[r10],r3 163 vst1.8 {d2},[r10],r3 165 vst1.8 {d3},[r10],r3 168 vst1.8 {d4},[r10],r3 170 vst1.8 {d5},[r10],r3 172 vst1.8 {d6},[r10],r3 174 vst1.8 {d7},[r10],r3 177 subeq r2,r10,r1 [all...] |
/external/valgrind/none/tests/arm/ |
H A D | v8memory_a.stdout.exp | 3 lda r6, [r10] with r10 = middle_of_block 16 00000000 r10 (xor, addr intreg #1) 18 ldab r9, [r10] with r10 = middle_of_block 31 00000000 r10 (xor, addr intreg #1) 33 ldah r3, [r10] with r10 = middle_of_block 46 00000000 r10 (xor, addr intreg #1) 50 stl r6, [r10] wit [all...] |
H A D | v8memory_t.stdout.exp | 3 lda r6, [r10] with r10 = middle_of_block 16 00000000 r10 (xor, addr intreg #1) 18 ldab r9, [r10] with r10 = middle_of_block 31 00000000 r10 (xor, addr intreg #1) 33 ldah r3, [r10] with r10 = middle_of_block 46 00000000 r10 (xor, addr intreg #1) 50 stl r6, [r10] wit [all...] |
/external/tremolo/Tremolo/ |
H A D | bitwiseARM.s | 45 STMFD r13!,{r10,r11,r14} 53 LDR r10,[r3] @ r10= ptr[0] 57 MOV r10,r10,LSR r14 @ r10= ptr[0]>>(32-bitsLeftInWord) 60 ORRLT r10,r10,r11,LSL r14 @ r10= Next 32 bits. 63 AND r0,r10,r1 [all...] |
H A D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 189 @ r10= wR 197 LDRB r6, [r10,#-1]! @ r6 = *--wR 227 LDMFD r12,{r8,r9,r10} @ r8 = step 229 @ r10= wR 238 LDRB r6, [r10,#-1]! @ r6 = *--wR 322 LDRB r10,[r5],r2 @ r10= T[0] T += step 327 MUL r9, r6, r10 @ r9 = s0*T[0] 331 MUL r12,r7, r10 [all...] |
/external/python/cpython2/Modules/_ctypes/libffi/src/x86/ |
H A D | unix64.S | 49 movq (%rsp), %r10 /* Load return address. */ 54 movq %r10, 24(%rax) /* Relocate return address. */ 57 movq %rdi, %r10 /* Save a copy of the register area. */ 62 movq (%r10), %rdi 63 movq 8(%r10), %rsi 64 movq 16(%r10), %rdx 65 movq 24(%r10), %rcx 66 movq 32(%r10), %r8 67 movq 40(%r10), %r9 73 leaq 176(%r10), [all...] |
H A D | darwin64.S | 48 movq (%rsp), %r10 /* Load return address. */ 53 movq %r10, 24(%rax) /* Relocate return address. */ 56 movq %rdi, %r10 /* Save a copy of the register area. */ 61 movq (%r10), %rdi 62 movq 8(%r10), %rsi 63 movq 16(%r10), %rdx 64 movq 24(%r10), %rcx 65 movq 32(%r10), %r8 66 movq 40(%r10), %r9 72 leaq 176(%r10), [all...] |
/external/vixl/test/aarch32/ |
H A D | test-assembler-rd-rn-rm-a32.cc | 100 {{{r2, r10, r12}, false, al, "r2 r10 r12", "r2_r10_r12"}, 108 {{r4, r5, r10}, false, al, "r4 r5 r10", "r4_r5_r10"}, 114 {{r10, r14, r3}, false, al, "r10 r14 r3", "r10_r14_r3"}, 116 {{r12, r10, r12}, false, al, "r12 r10 r12", "r12_r10_r12"}, 117 {{r4, r10, r12}, false, al, "r4 r10 r1 [all...] |
H A D | test-assembler-rd-rn-rm-t32.cc | 100 {{{r2, r10, r12}, false, al, "r2 r10 r12", "r2_r10_r12"}, 108 {{r4, r5, r10}, false, al, "r4 r5 r10", "r4_r5_r10"}, 114 {{r10, r14, r3}, false, al, "r10 r14 r3", "r10_r14_r3"}, 116 {{r12, r10, r12}, false, al, "r12 r10 r12", "r12_r10_r12"}, 117 {{r4, r10, r12}, false, al, "r4 r10 r1 [all...] |
H A D | test-macro-assembler-cond-rd-rn-a32.cc | 96 {{lt, r11, r10}, "lt, r11, r10", "lt_r11_r10"}, 97 {{vs, r10, r8}, "vs, r10, r8", "vs_r10_r8"}, 106 {{cc, r4, r10}, "cc, r4, r10", "cc_r4_r10"}, 111 {{ls, r10, r10}, "ls, r10, r10", "ls_r10_r1 [all...] |
H A D | test-macro-assembler-cond-rd-rn-t32.cc | 96 {{lt, r11, r10}, "lt, r11, r10", "lt_r11_r10"}, 97 {{vs, r10, r8}, "vs, r10, r8", "vs_r10_r8"}, 106 {{cc, r4, r10}, "cc, r4, r10", "cc_r4_r10"}, 111 {{ls, r10, r10}, "ls, r10, r10", "ls_r10_r1 [all...] |
H A D | test-assembler-cond-rd-rn-rm-t32.cc | 154 {{al, r1, r2, r10}, false, al, "al r1 r2 r10", "al_r1_r2_r10"}, 157 {{al, r6, r9, r10}, false, al, "al r6 r9 r10", "al_r6_r9_r10"}, 165 {{al, r10, r3, r13}, false, al, "al r10 r3 r13", "al_r10_r3_r13"}, 166 {{al, r10, r10, r2}, false, al, "al r10 r10 r [all...] |
/external/boringssl/ios-arm/crypto/chacha/ |
H A D | chacha-armv4.S | 66 ldmia r3,{r4,r5,r6,r7,r8,r9,r10,r11} @ load key 68 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11} @ copy key 70 str r10,[sp,#4*(16+10)] @ off-load "rx" 83 ldr r10, [sp,#4*(13)] 95 mov r10,r10,ror#16 97 eor r10,r10,r1,ror#16 100 add r9,r9,r10 107 mov r10,r1 [all...] |
/external/boringssl/linux-arm/crypto/chacha/ |
H A D | chacha-armv4.S | 65 ldmia r3,{r4,r5,r6,r7,r8,r9,r10,r11} @ load key 67 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11} @ copy key 69 str r10,[sp,#4*(16+10)] @ off-load "rx" 82 ldr r10, [sp,#4*(13)] 94 mov r10,r10,ror#16 96 eor r10,r10,r1,ror#16 99 add r9,r9,r10 106 mov r10,r1 [all...] |
/external/boringssl/ios-arm/crypto/fipsmodule/ |
H A D | sha1-armv4-large.S | 32 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11,r12,lr} 44 ldrb r10,[r1,#2] 49 orr r9,r9,r10,lsl#8 50 eor r10,r5,r6 @ F_xx_xx 57 eor r10,r5,r6 @ F_xx_xx 63 and r10,r4,r10,ror#2 65 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 67 add r7,r7,r10 [all...] |
/external/boringssl/linux-arm/crypto/fipsmodule/ |
H A D | sha1-armv4-large.S | 31 stmdb sp!,{r4,r5,r6,r7,r8,r9,r10,r11,r12,lr} 43 ldrb r10,[r1,#2] 48 orr r9,r9,r10,lsl#8 49 eor r10,r5,r6 @ F_xx_xx 56 eor r10,r5,r6 @ F_xx_xx 62 and r10,r4,r10,ror#2 64 eor r10,r10,r6,ror#2 @ F_00_19(B,C,D) 66 add r7,r7,r10 [all...] |
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
H A D | x86_64-mont5.asm | 52 lea r10,[((-280))+r9*8+rsp] 54 and r10,-1024 64 sub r11,r10 66 lea rsp,[r11*1+r10] 68 cmp rsp,r10 75 cmp rsp,r10 79 lea r10,[$L$inc] 85 movdqa xmm0,XMMWORD[r10] 86 movdqa xmm1,XMMWORD[16+r10] 87 lea r10,[((2 [all...] |
/external/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_linux_x86_64.S | 20 mov %rcx, %r10
|
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
H A D | x86_64-mont.S | 43 leaq -16(%rsp,%r9,8),%r10 45 andq $-1024,%r10 55 subq %r10,%r11 57 leaq (%r10,%r11,1),%rsp 59 cmpq %r10,%rsp 67 cmpq %r10,%rsp 84 movq %rax,%r10 87 imulq %r10,%rbp 91 addq %rax,%r10 105 movq %r10, [all...] |
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
H A D | x86_64-mont.S | 42 leaq -16(%rsp,%r9,8),%r10 44 andq $-1024,%r10 54 subq %r10,%r11 56 leaq (%r10,%r11,1),%rsp 58 cmpq %r10,%rsp 66 cmpq %r10,%rsp 83 movq %rax,%r10 86 imulq %r10,%rbp 90 addq %rax,%r10 104 movq %r10, [all...] |