/external/openssl/crypto/sha/asm/ |
H A D | sha1-armv4-large.s | 23 ldrb r9,[r1,#3] 27 orr r9,r9,r10,lsl#8 29 orr r9,r9,r11,lsl#16 31 orr r9,r9,r12,lsl#24 33 ldr r9,[r1],#4 @ handles unaligned 38 rev r9,r9 [all...] |
H A D | sha512-ia64.pl | 148 add r9=1*$SZ,ctx 157 $LDW B_=[r9],4*$SZ 163 $LDW F_=[r9] 200 { .mmi; add r9=1-$SZ,input 203 { .mmi; ld1 r9=[r9] 210 dep X[15]=X[15],r9,8,8 216 and r9=A,C 222 xor T2=T2,r9 };; 227 add r9 [all...] |
/external/openssl/crypto/bn/asm/ |
H A D | ppc.pl | 232 #.set r9,9 275 # Freely use registers r5,r6,r7,r8,r9,r10,r11 as follows: 279 # r9,r10, r11 are the equivalents of c1,c2, c3. 288 $UMULL r9,r5,r5 294 $ST r9,`0*$BNSZ`(r3) # r[0]=c1; 302 addze r9,r0 # catch carry if any. 303 # r9= r0(=0) and carry 307 addze r9,r9 314 adde r9,r [all...] |
/external/libvpx/vp8/common/arm/armv6/ |
H A D | recon_v6.asm | 41 stmdb sp!, {r4 - r9, lr} 49 pkhtb r9, r7, r6, asr #16 ; 3 | 1 52 uxtab16 r9, r9, r4, ror #8 ; 3 | 1 + 0 | 3 | 2 | 1 55 usat16 r9, #8, r9 57 orr r8, r8, r9, lsl #8 69 pkhtb r9, r7, r6, asr #16 ; 3 | 1 72 uxtab16 r9, r9, r [all...] |
H A D | loopfilter_v6.asm | 74 ldr r9, [src], pstep ; p3 89 uqsub8 r6, r9, r10 ; p3 - p2 90 uqsub8 r7, r10, r9 ; p2 - p3 101 ldr r9, [src], pstep ; q0 110 uqsub8 r11, r12, r9 ; p0 - q0 111 uqsub8 r12, r9, r12 ; q0 - p0 118 uqsub8 r7, r9, r10 ; q0 - q1 120 uqsub8 r6, r10, r9 ; q1 - q0 122 uqsub8 r9, r11, r10 ; q2 - q1 129 orr r10, r9, r1 [all...] |
H A D | simpleloopfilter_v6.asm | 73 ldr r9, [sp, #40] ; count for 8-in-parallel 75 mov r9, r9, lsl #1 ; double the count. we're doing 4 at a time 132 subs r9, r9, #1 155 ; load soure data to r7, r8, r9, r10 178 pkhbt r9, r3, r4, lsl #16 181 ;transpose r7, r8, r9, r10 to r3, r4, r5, r6 182 TRANSPOSE_MATRIX r7, r8, r9, r10, r3, r4, r5, r6 186 uqsub8 r9, r [all...] |
/external/libyuv/files/source/ |
H A D | rotate_neon.s | 108 push {r4,r8,r9,lr} 119 mov r9, r0 121 vld1.8 {d0}, [r9], r1 122 vld1.8 {d1}, [r9], r1 123 vld1.8 {d2}, [r9], r1 124 vld1.8 {d3}, [r9], r1 125 vld1.8 {d4}, [r9], r1 126 vld1.8 {d5}, [r9], r1 127 vld1.8 {d6}, [r9], r1 128 vld1.8 {d7}, [r9] [all...] |
/external/valgrind/main/coregrind/m_syswrap/ |
H A D | syscall-s390x-linux.S | 96 we clobber (r6-r9) */ 97 stmg %r2,%r9, SP_R2(%r15) 108 lg %r9, SP_R3(%r15) /* guest state --> r9 */ 109 lg %r2, OFFSET_s390x_r2(%r9) /* guest r2 --> real r2 */ 110 lg %r3, OFFSET_s390x_r3(%r9) /* guest r3 --> real r3 */ 111 lg %r4, OFFSET_s390x_r4(%r9) /* guest r4 --> real r4 */ 112 lg %r5, OFFSET_s390x_r5(%r9) /* guest r5 --> real r5 */ 113 lg %r6, OFFSET_s390x_r6(%r9) /* guest r6 --> real r6 */ 114 lg %r7, OFFSET_s390x_r7(%r9) /* gues [all...] |
/external/tremolo/Tremolo/ |
H A D | mdctLARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 188 @ r9 = wL 198 LDRB r11,[r9],#1 @ r11= *wL++ 227 LDMFD r12,{r8,r9,r10} @ r8 = step 228 @ r9 = wL 237 LDRB r11,[r9],#1 @ r11= *wL++ 327 MUL r9, r6, r10 @ r9 = s0*T[0] 329 MLA r9, r7, r11,r9 [all...] |
H A D | mdctARM.s | 187 LDMFD r12,{r8,r9,r10} @ r8 = step 188 @ r9 = wL 196 LDR r11,[r9],#4 @ r11= *wL++ 226 LDMFD r12,{r8,r9,r10} @ r8 = step 227 @ r9 = wL 235 LDR r11,[r9],#4 @ r11= *wL++ 324 SMULL r8, r9, r7, r11 @ (r8, r9) = s2*T[1] 327 SMLAL r8, r9, r6, r10 @ (r8, r9) [all...] |
/external/libvpx/vp8/decoder/arm/armv6/ |
H A D | dequant_idct_v6.asm | 68 smulwt r9, r3, r6 72 pkhbt r7, r7, r9, lsl #16 77 smulwb r9, r3, r12 80 pkhbt r9, r9, r11, lsl #16 83 uadd16 r7, r12, r9 88 uadd16 r9, r10, r6 96 str r9, [r1], #4 105 ldr r9, [r0], #4 115 pkhbt lr, r9, r [all...] |
H A D | dequant_dc_idct_v6.asm | 90 smulwt r9, r3, r6 94 pkhbt r7, r7, r9, lsl #16 99 smulwb r9, r3, r12 102 pkhbt r9, r9, r11, lsl #16 105 uadd16 r7, r12, r9 110 uadd16 r9, r10, r6 118 str r9, [r1], #4 127 ldr r9, [r0], #4 137 pkhbt lr, r9, r [all...] |
H A D | dequantize_v6.asm | 21 stmdb sp!, {r4-r9, lr} 33 smulbb r9, r5, r6 45 strh r9, [r2], #2 46 smulbb r9, r5, r6 60 strh r9, [r2], #2 66 ldmia sp!, {r4-r9, pc}
|
/external/llvm/test/MC/ARM/ |
H A D | arm_fixups.s | 8 mov r9, :lower16:(_foo) 9 movw r9, :lower16:(_foo) 10 movt r9, :upper16:(_foo) 12 @ CHECK: movw r9, :lower16:_foo @ encoding: [A,0x90'A',0b0000AAAA,0xe3] 14 @ CHECK: movw r9, :lower16:_foo @ encoding: [A,0x90'A',0b0000AAAA,0xe3] 16 @ CHECK: movt r9, :upper16:_foo @ encoding: [A,0x90'A',0b0100AAAA,0xe3]
|
/external/libvpx/vp8/encoder/arm/armv6/ |
H A D | vp8_subtract_armv6.asm | 29 stmfd sp!, {r4-r9} 40 mov r9, #4 ; loop count 55 subs r9, r9, #1 ; decrement loop counter 66 ldmfd sp!, {r4-r9} 93 uxtb16 r9, r7 ; [p2 | p0] (A) 97 usub16 r6, r8, r9 ; [d2 | d0] (A) 104 pkhtb r9, r7, r6, asr #16 ; [d3 | d2] (A) 108 str r9, [r0], #4 ; diff (A) 110 uxtb16 r9, r1 [all...] |
H A D | vp8_fast_fdct4x4_armv6.asm | 49 ldrd r8, r9, [r0] ; [i5 | i4] [i7 | i6] 57 ror r9, r9, #16 ; [i6 | i7] 59 qadd16 r6, r8, r9 ; [i5+i6 | i4+i7] = [b1 | a1] without shift 60 qsub16 r7, r8, r9 ; [i5-i6 | i4-i7] = [c1 | d1] without shift 68 smuad r9, r6, lr ; o4 = (i5+i6)*8 + (i4+i7)*8 76 pkhbt r9, r9, r6, lsl #4 ; [o5 | o4], keep in register for PART 2 134 qadd16 r5, r9, r2 ; b1 = [i5+i9 | i4+i8] 135 qsub16 r6, r9, r [all...] |
/external/aac/libFDK/src/arm/ |
H A D | dct_arm.cpp | 110 r9 accuX 112 PUSH {r4-r9} 121 SMULWT r9, r5, r8 // accuX = accu2*val_tw.l local 123 RSB r9, r9, #0 // accuX =-accu2*val_tw.l local 125 SMLAWB r4, r4, r8, r9 // accu1 = accu1*val_tw.h - accu2*val_tw.l 131 SMULWB r9, r7, r8 // accuX = accu4*val_tw.h local 133 RSB r9, r9, #0 // accuX =-accu4*val_tw.h local 135 SMLAWT r6, r6, r8, r9 // accu 147 SMULWT r9, r5, r8 // accuX = accu2*val_tw.l local 149 RSB r9, r9, #0 // accuX =-accu2*val_tw.l local 157 SMULWB r9, r7, r8 // accuX = accu4*val_tw.h local 159 RSB r9, r9, #0 // accuX =-accu4*val_tw.h local 298 SMULWT r9, r5, r8 // accuX = (-accu2)*val_tw.l local 300 RSB r9, r9, #0 // accuX = -(-accu2)*val_tw.l local 301 SMLAWB r9, r4, r8, r9 // accuX = accu1*val_tw.h-(-accu2)*val_tw.l local 307 STR r9, [r2], #4 // *pDat_0++ = accu1 (accuX) local 321 SMULWT r9, r5, r8 // accuX = (-accu2)*val_tw.l local 323 RSB r9, r9, #0 // accuX = -(-accu2)*val_tw.l local 324 SMLAWB r9, r4, r8, r9 // accuX = accu1*val_tw.h-(-accu2)*val_tw.l local 330 STR r9, [r2], #4 // *pDat_0++ = accu1 (accuX) local [all...] |
/external/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframe_func_neon.asm | 31 ldr r9, [r1, #yv12_buffer_config_u_buffer] ;srcptr1 43 str r9, [sp, #4] 52 mov r9, r3 69 vst1.8 {q0, q1}, [r9]! 71 vst1.8 {q2, q3}, [r9]! 73 vst1.8 {q4, q5}, [r9]! 75 vst1.8 {q6, q7}, [r9]! 110 mov r9, r3 123 vst1.8 {q0, q1}, [r9]! 125 vst1.8 {q2, q3}, [r9]! [all...] |
/external/libvpx/vp8/encoder/arm/armv5te/ |
H A D | boolhuff_armv5te.asm | 45 push {r4-r9, lr} 83 mov r9, #0 84 strb r9, [r7, r4] ; w->buffer[x] =(unsigned char)0 94 ldrb r9, [r7, r4] ; w->buffer[x] 95 add r9, r9, #1 96 strb r9, [r7, r4] ; w->buffer[x] + 1 99 ldr r9, [r0, #vp8_writer_buffer] 108 strb r7, [r9, r4] ; w->buffer[w->pos++] 116 pop {r4-r9, p [all...] |
/external/v8/src/arm/ |
H A D | codegen-arm.cc | 91 r9, 126 __ AllocateInNewSpace(lr, r6, r7, r9, &gc_required, NO_ALLOCATION_FLAGS); 129 __ LoadRoot(r9, Heap::kFixedDoubleArrayMapRootIndex); 131 __ str(r9, MemOperand(r6, HeapObject::kMapOffset)); 138 r9, 149 r9, 175 r9, 189 __ ldr(r9, MemOperand(r3, 4, PostIndex)); 190 // r9: current element 191 __ UntagAndJumpIfNotSmi(r9, r [all...] |
/external/clang/test/CXX/except/except.spec/ |
H A D | p3.cpp | 40 extern void (*r9)() throw(int, int); 41 extern void (*r9)() throw(int, int);
|
/external/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_rtl_amd64.S | 25 push %r9 27 .cfi_rel_offset %r9, 0 55 pop %r9 76 .cfi_restore %r9 104 push %r9 106 .cfi_rel_offset %r9, 0 134 pop %r9 155 .cfi_restore %r9
|
/external/valgrind/main/none/tests/amd64/ |
H A D | sbbmisc.stdout.exp | 9 r9 = -14982 -14981
|
/external/valgrind/main/none/tests/x86/ |
H A D | sbbmisc.stdout.exp | 9 r9 = -14982 -14981
|
/external/openssl/crypto/aes/asm/ |
H A D | aes-armv4.s | 232 and r9,lr,r0,lsr#16 239 ldr r6,[r10,r9,lsl#2] @ Te1[s0>>16] 240 and r9,lr,r1,lsr#8 246 ldr r9,[r10,r9,lsl#2] @ Te2[s1>>8] 252 eor r6,r6,r9,ror#8 253 and r9,lr,r2 259 ldr r9,[r10,r9,lsl#2] @ Te3[s2>>0] 265 eor r6,r6,r9,ro [all...] |