/external/libvpx/vp8/common/ppc/ |
H A D | platform_altivec.asm | 25 ;# r3 context_ptr 28 W v20, r3 29 W v21, r3 30 W v22, r3 31 W v23, r3 32 W v24, r3 33 W v25, r3 34 W v26, r3 35 W v27, r3 36 W v28, r3 [all...] |
/external/openssl/crypto/sha/asm/ |
H A D | sha256-armv4.s | 28 sub r3,pc,#8 @ sha256_block_data_order 32 sub r14,r3,#256 @ K256 35 ldrb r3,[r1,#3] @ 0 39 orr r3,r3,r12,lsl#8 40 orr r3,r3,r2,lsl#16 41 orr r3,r3,r0,lsl#24 44 str r3,[s [all...] |
H A D | sha512-armv4.s | 51 sub r3,pc,#8 @ sha512_block_data_order 54 sub r14,r3,#640 @ K512 70 ldr r3,[r0,#8+4] 76 str r3,[sp,#8+0] 82 ldr r3,[r0,#40+4] 84 str r3,[sp,#40+0] 88 ldrb r3,[r1,#7] 94 orr r3,r3,r9,lsl#8 96 orr r3,r [all...] |
/external/skia/src/core/asm/ |
H A D | s32a_d565_opaque.S | 43 andeq r3, r6, #63488 // 0xf800 45 orreq r3, r3, r2 46 orreq r3, r3, r6, lsr #27 47 streqh r3, [r0], #2 // *dst = r3; dst++ 50 mov r3, r1, lsl #16 52 mov r5, r3, lsr #24 53 ldrh r3, [r [all...] |
/external/openssl/crypto/ |
H A D | ppccpuid.pl | 35 mr r3,r1 50 Loop: lwarx r5,0,r3 52 stwcx. r0,0,r3 54 $SIGNX r3,r0 60 mftb r3 73 stb r0,0(r3) 74 addi r3,r3,1 77 Lot: andi. r5,r3,3 79 stb r0,0(r3) [all...] |
H A D | s390xcpuid.S | 30 ar %r0,%r3 44 xgr %r3,%r3 63 clgr %r3,%r4 65 clgr %r3,%r0 70 brctg %r3,.Little 77 brctg %r3,.Lot 79 srlg %r4,%r3,3 84 ngr %r3,%r4
|
/external/libvpx/vp8/decoder/arm/ |
H A D | detokenize.asm | 44 mov r3, #16 ; stop 48 addeq r3, r3, #8 ; stop = 24 54 str r3, [sp, #l_stop] 69 ldr r3, [r9, #detok_ptr_block2leftabove] 72 ldrb r12, [r3, r11]! ; block2left[i] 73 ldrb r3, [r3, #c_block2above_offset]; block2above[i] 80 ldrb r3, [r2, r3]! ; *( [all...] |
/external/libvpx/vp8/encoder/ppc/ |
H A D | variance_altivec.asm | 75 load_aligned_16 v4, r3, r10 79 add r3, r3, r4 90 lwz r3, 12(r1) 96 stw r3, 0(r8) ;# sum 100 mullw r3, r3, r3 ;# sum*sum 101 srawi r3, r3, \D [all...] |
H A D | sad_altivec.asm | 59 lvx v4, 0, r3 64 add r3, r3, r4 75 lvx v9, 0, r3 81 add r3, r3, r4 89 lvx v4, 0, r3 97 add r3, r3, r4 109 lwz r3, 1 [all...] |
H A D | variance_subpixel_altivec.asm | 64 ;# r3 src_ptr 83 lvsl v17, 0, r3 ;# permutate value for alignment 86 lvx v21, 0, r3 87 lvx v22, r10, r3 90 add r3, r3, r4 152 lwz r3, 12(r1) 159 mullw r3, r3, r3 ;# su [all...] |
/external/jpeg/ |
H A D | armv6_idct.S | 84 stm r5, {r2, r3, r4} 95 ldrsh r3, [r14, #48] 107 orreqs r8, r3, r5 131 // r3 = q[3] * c[3] + r5; 135 mla r3, r11, r3, r5 143 // r1 = r1 - r3; 144 // r5 = r5 * 2 - r3 = -(r3 - r5 * 2); 145 // r3 [all...] |
/external/libffi/src/powerpc/ |
H A D | aix_closure.S | 30 .set r3,3 120 stw r3, 200(r1) 146 mr r3,r11 161 /* now r3 contains the return type */ 169 slwi r3,r3,2 /* now multiply return type by 4 */ 170 lwzx r3,r4,r3 /* get the contents of that table value */ 171 add r3,r3,r [all...] |
H A D | ppc_closure.S | 48 stw %r3, 16(%r1) 71 mr %r3,%r11 90 # now r3 contains the return type 98 slwi %r3,%r3,4 # now multiply return type by 16 101 add %r3,%r3,%r4 # add contents of table to table address 102 mtctr %r3 118 lwz %r3,112+0(%r1) 143 lbz %r3,11 [all...] |
H A D | darwin_closure.S | 68 stw r3,200(r1) 94 mr r3,r11 108 /* Now r3 contains the return type 117 slwi r3,r3,4 /* Now multiply return type by 16. */ 118 add r3,r3,r4 /* Add contents of table to table address. */ 119 mtctr r3 143 lwz r3,0(r5) 171 lbz r3, [all...] |
H A D | linux64_closure.S | 47 std %r3, 48(%r1) 81 mr %r3, %r11 97 # now r3 contains the return type 104 sldi %r3, %r3, 4 # now multiply return type by 16 107 add %r3, %r3, %r4 # add contents of table to table address 108 mtctr %r3 123 lwa %r3, 112+4(%r1) 143 lbz %r3, 11 [all...] |
/external/libvpx/vp8/decoder/arm/armv5/ |
H A D | dequantize_v5.asm | 23 ldr r3, [q] 28 smulbb lr, r3, r4 29 smultt r12, r3, r4 31 ldr r3, [q, #4] 37 smulbb lr, r3, r4 38 smultt r12, r3, r4 41 ldrne r3, [q, #4]
|
/external/libvpx/vp8/common/arm/armv6/ |
H A D | simpleloopfilter_v6.asm | 54 ;r3 const char *limit, 66 ldr r12, [r3] ; limit 67 ldr r3, [src, -pstep, lsl #1] ; p1 82 uqsub8 r7, r3, r6 ; p1 - q1 83 uqsub8 r8, r6, r3 ; q1 - p1 99 eor r3, r3, r2 ; p1 offset to convert to a signed value 104 qsub8 r3, r3, r6 ; vp8_filter = p1 - q1 106 qadd8 r3, r [all...] |
H A D | iwalsh_v6.asm | 26 ldr r3, [r0], #4 ; [3 | 2] 44 qadd16 r10, r3, r9 ; a1 [3+15 | 2+14] 47 qsub16 lr, r3, r9 ; d1 [3-15 | 2-14] 49 qadd16 r3, r10, r11 ; a1 + b1 [3 | 2] 56 qsubaddx r10, r2, r3 ; [c1|a1] [1-2 | 0+3] 57 qaddsubx r11, r2, r3 ; [b1|d1] [1+2 | 0-3] 62 qaddsubx r3, r11, r10 ; [a2|d2] [b1+a1 | d1-c1] 68 qadd16 r3, r3, r10 ; [a2+3|d2+3] 73 pkhtb r12, r12, r3, as [all...] |
/external/tremolo/Tremolo/ |
H A D | bitwiseARM.s | 46 LDMIA r0,{r2,r3,r12} 48 @ r3 = ptr 53 LDR r10,[r3] @ r10= ptr[0] 56 LDRLT r11,[r3,#4]! @ r11= ptr[1] 78 LDR r10,[r3],#4 @ r10= ptr[0] 79 LDRLT r6,[r3] @ r6 = ptr[1] 162 LDMIA r0,{r2,r3,r12} 164 @ r3 = ptr 170 ADDLE r3,r3,# [all...] |
/external/libvpx/vp8/encoder/arm/neon/ |
H A D | subtract_neon.asm | 26 vld1.8 {d0}, [r1], r3 ;load src 28 vld1.8 {d2}, [r1], r3 30 vld1.8 {d4}, [r1], r3 32 vld1.8 {d6}, [r1], r3 56 vld1.8 {q0}, [r1], r3 ;load src 58 vld1.8 {q2}, [r1], r3 60 vld1.8 {q4}, [r1], r3 62 vld1.8 {q6}, [r1], r3 96 add r3, r3, #25 [all...] |
/external/v8/src/arm/ |
H A D | ic-arm.cc | 59 // r3 - used as temporary and to hold the capacity of the property 70 __ ldr(r3, FieldMemOperand(t0, Map::kInstanceAttributesOffset)); 71 __ tst(r3, Operand(1 << (Map::kHasNamedInterceptor + (3 * 8)))); 76 __ ldrb(r3, FieldMemOperand(t0, Map::kInstanceTypeOffset)); 77 __ cmp(r3, Operand(JS_GLOBAL_PROXY_TYPE)); 82 __ cmp(r3, Operand(JS_GLOBAL_OBJECT_TYPE)); 84 __ cmp(r3, Operand(JS_BUILTINS_OBJECT_TYPE)); 89 __ ldr(r3, FieldMemOperand(t0, HeapObject::kMapOffset)); 91 __ cmp(r3, ip); 97 __ ldr(r3, FieldMemOperan [all...] |
/external/libffi/src/m32r/ |
H A D | sysv.S | 51 push r3 72 ld r3, @(12,sp) 93 ld r3, @(12,sp) 100 bne r3, r4, return_double 107 bne r3, r4, epilogue 115 pop r3
|
/external/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframe_func_neon.asm | 40 ldr r3, [r1, #yv12_buffer_config_y_buffer] ;dstptr1 52 mov r9, r3 54 add r11, r3, r7 82 add r3, r3, r7, lsl #1 89 ldr r3, [r1, #yv12_buffer_config_y_buffer] ;dstptr1 95 ldr r3, [sp, #4] ;dstptr1 110 mov r9, r3 112 add r11, r3, r7 132 add r3, r [all...] |
H A D | vp8_vpxyv12_copysrcframe_func_neon.asm | 36 ldr r3, [r1, #yv12_buffer_config_y_buffer] ;dstptr1 39 add r11, r3, r7 ;second row dst 62 vst1.8 {q0, q1}, [r3]! 64 vst1.8 {q2, q3}, [r3]! 66 vst1.8 {q8, q9}, [r3]! 68 vst1.8 {q10, q11}, [r3]! 80 vst1.8 {d0}, [r3]! 90 strb r8, [r3], #1 98 add r3, r3, r [all...] |
/external/libvpx/vp8/common/arm/neon/ |
H A D | buildintrapredictorsmby_neon.asm | 23 ; r3 int mode 30 cmp r3, #0 32 cmp r3, #1 34 cmp r3, #2 36 cmp r3, #3 79 ldrb r3, [r0], r2 84 add r12, r12, r3 89 ldrb r3, [r0], r2 94 add r12, r12, r3 99 ldrb r3, [r [all...] |