/frameworks/av/media/libstagefright/codecs/on2/h264dec/omxdl/arm_neon/vc/m4p10/src_gcc/ |
H A D | armVCM4P10_InterpolateLuma_DiagCopy_unsafe_s.S | 38 LDR r5,[r0,#4] 42 UQSUB16 r5,r5,r6 46 USAT16 r5,#13,r5 50 AND r5,r12,r5,LSR #5 53 ORR r10,r4,r5,LSL #8 72 LDR r5,[r0,#4] 76 UQSUB16 r5,r [all...] |
H A D | armVCM4P10_Average_4x_Align_unsafe_s.S | 39 UHSUB8 r5,r10,r12 41 EOR r5,r5,r7 42 STR r5,[r2],r3 50 UHSUB8 r5,r10,r12 53 EOR r5,r5,r7 54 STR r5,[r2],r3 69 LDR r5,[r0,#4] 76 ORR r11,r11,r5,LS [all...] |
H A D | armVCM4P10_UnpackBlock4x4_s.S | 35 MOV r5,#0 37 STRD r4,r5,[r1,#0] 38 STRD r4,r5,[r1,#8] 39 STRD r4,r5,[r1,#0x10] 40 STRD r4,r5,[r1,#0x18] 43 LDRNESB r5,[r2,#1] 47 ORRNE r4,r4,r5,LSL #8
|
H A D | armVCM4P10_InterpolateLuma_Copy_unsafe_s.S | 42 LDR r5,[r0],r1 45 STR r5,[r2],r3 51 LDR r5,[r0,#4] 56 ORR r4,r4,r5,LSL #24 60 LDR r5,[r0,#4] 66 ORR r4,r4,r5,LSL #24 73 LDR r5,[r0,#4] 78 ORR r4,r4,r5,LSL #16 83 LDR r5,[r0,#4] 88 ORR r4,r4,r5,LS [all...] |
H A D | omxVCM4P10_TransformDequantChromaDCFromPair_s.S | 39 ldrnesb r5, [r9, #1] 43 orrne r4, r4, r5, lsl #8 50 ldr r5, .LarmVCM4P10_QPDivTable 51 P0: add r5, pc 56 ldrsb r9, [r5, r2] 58 sadd16 r5, r3, r4 61 vmov d0, r5, r6
|
H A D | omxVCM4P10_PredictIntra_4x4_s.S | 44 LDRD r4,r5,[sp,#0x50] 55 ADD r11,r3,r5 56 ADD r12,r5,r5 64 ADD r11,r3,r5 65 ADD r12,r5,r5 89 ADD r11,r3,r5 90 ADD r12,r5,r5 [all...] |
H A D | omxVCM4P10_DeblockLuma_I.S | 38 LDR r5,[sp,#0x2c] 49 CMP r5,#0 51 TST r5,#3 61 STR r5,[sp,#4] 68 ADD r3,r5,#0x10
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV7/ |
H A D | convolve_neon.s | 41 ADD r5, r3, #1 @ i = n + 1 45 SUB r5, r5, #1 50 CMP r5, #0 57 SUBS r5, r5, #4 63 VMOV.S32 r5, D20[0] 64 ADD r5, r5, r8 65 ADD r5, r1 [all...] |
H A D | cor_h_vec_neon.s | 31 @r5 ---- cor_1[] 48 MOV r5, #0 @L_sum1 = 0 59 MLA r5, r12, r8, r5 65 MLA r5, r12, r14, r5 67 MOV r5, r5, LSL #2 @L_sum1 = (L_sum1 << 2) 69 ADD r9, r5, r14 70 MOV r5, r [all...] |
H A D | Norm_Corr_neon.s | 31 @ r5 --- t_max 59 ADD r5, r0, r11, LSL #1 @get the &exc[k] 63 MOV r0, r5 165 VMOV.S32 r5, D22[0] 167 @r5 --- L_tmp, r6 --- L_tmp1 169 ADD r5, r10, r5, LSL #1 @L_tmp = (L_tmp << 1) + 1 172 CLZ r10, r5 173 CMP r5, #0 174 RSBLT r11, r5, # [all...] |
H A D | syn_filt_neon.s | 39 MOV r5, r13 @ copy yy = y_buf address 46 VST1.S16 {D0, D1, D2, D3}, [r5]! @store 16 mem[] to *yy 48 LDRSH r5, [r0], #2 @ load a[0] 50 MOV r5, r5, ASR #1 @ a0 = a[0] >> 1 51 VMOV.S16 D8[0], r5 67 MUL r12, r6, r5 @ L_tmp = x[i] * a0 95 ADD r5, r13, #160 @ yy[64] address 96 VLD1.S16 {D0, D1, D2, D3}, [r5]!
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/src/asm/ARMV5E/ |
H A D | convolve_opt.s | 40 ADD r5, r3, #1 @ i = n + 1 44 SUB r5, r5, #1 48 CMP r5, #0 61 SUBS r5, r5, #4 68 ADD r5, r11, r8, LSL #1 69 MOV r5, r5, LSR #16 @extract_h(s) 71 STRH r5, [r [all...] |
H A D | cor_h_vec_opt.s | 30 @r5 ---- cor_1[] 47 MOV r5, #0 @L_sum1 = 0 58 MLA r5, r12, r8, r5 64 MLA r5, r12, r14, r5 66 MOV r5, r5, LSL #2 @L_sum1 = (L_sum1 << 2) 68 ADD r9, r5, r14 69 MOV r5, r [all...] |
H A D | scale_sig_opt.s | 44 LDRSH r5, [r4] @load x[i] 45 MOV r12, r5, LSL r10 46 TEQ r5, r12, ASR r10 47 EORNE r12, r8, r5, ASR #31 57 LDRSH r5, [r4] @load x[i] 58 MOV r6, r5, LSL #16 @L_tmp = x[i] << 16 59 MOV r5, r6, ASR r7 @L_tmp >>= exp 60 QADD r11, r5, r9
|
H A D | Norm_Corr_opt.s | 31 @ r5 --- t_max 59 ADD r5, r0, r11, LSL #1 @get the &exc[k] 63 MOV r0, r5 72 MOV r5, #64 89 SUBS r5, r5, #8 104 MOV r5, #0 @L_tmp = 0 116 SMLABB r5, r10, r11, r5 @L_tmp += xn[i] * excf[i] 117 SMLATT r5, r1 [all...] |
H A D | syn_filt_opt.s | 40 MOV r5, r13 @ copy yy = y_buf address 56 STRH r6, [r5], #2 57 STRH r7, [r5], #2 58 STRH r8, [r5], #2 59 STRH r9, [r5], #2 60 STRH r10, [r5], #2 61 STRH r11, [r5], #2 62 STRH r12, [r5], #2 63 STRH r14, [r5], #2 74 STRH r6, [r5], # [all...] |
/frameworks/av/media/libstagefright/codecs/mp3dec/src/asm/ |
H A D | pvmp3_mdct_18_gcc.s | 51 mov r5,r0 92 mov r0,r5 @@ r0 = vec 94 add r0,r5,#0x24 @@ r0 = &vec[9] 97 ldr r0,[r5,#0x20] 98 ldr r2,[r5,#0x40] 99 str r0,[r5,#0x40] 100 ldr r0,[r5,#0x1c] 101 ldr r3,[r5,#0x38] 102 str r0,[r5,#0x38] 103 ldr r1,[r5,# [all...] |
H A D | pvmp3_polyphase_filter_window_gcc.s | 69 ldr r5,[r3] 73 smlal r2,r9,lr,r5 77 smlal r5,r11,r2,r5 78 smull r6,r5,r2,r6 79 sub r9,r9,r5 80 ldr r5,[r1,#8] 83 smlal r6,r9,r5,r7 84 smull r6,r2,r5,r8 85 ldr r5,[r [all...] |
/frameworks/native/opengl/libagl/ |
H A D | iterators.S | 51 stmfd sp!, {r4, r5, r6, r7, r8, lr} 61 ldr r5, [r0, #0] 63 smull r8, lr, r4, r5 64 ldr r5, [r0, #8] 67 smlal r8, lr, r3, r5 76 umull r6, r5, r3, r6 78 mla r7, r3, r7, r5 88 ldmfd sp!, {r4, r5, r6, r7, r8, pc}
|
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV5E/ |
H A D | band_nrg_v5.s | 34 mov r5, r2, asr #16 36 cmp r5, #0 73 cmp r4, r5 98 mov r5, r4, lsl #1 100 ldrsh r10, [r2, r5] 101 add r5, r2, r5 103 ldrsh r11, [r5, #2] 120 add r5, r8, r9 124 smull r5, r [all...] |
H A D | R4R8First_v5.s | 47 add r3, r5, r7 50 sub r5, r5, r7 58 add r2, r0, r5 61 sub r0, r0, r5 102 add r3, r5, r7 @ i2 = buf[5] + buf[7]@ 105 sub r5, r5, r7 @ i3 = buf[5] - buf[7]@ 110 sub r8, r8, r2 @ r5 = (r0 - r2) >> 1@ 113 sub r2, r0, r5 [all...] |
H A D | Radix4FFT_v5.s | 65 smulwb r5, r10, r8 @ L_mpy_wx(sinx, t0) 70 sub r3, r3, r5 @ r3 = L_mpy_wx(cosx, t1) - L_mpy_wx(sinx, t0)@ 82 ldrd r10, [r14, #0] @ r4 = xptr[0]@ r5 = xptr[1]@ 86 smulwt r5, r11, r8 @ L_mpy_wx(cosx, t1) 92 sub r5, r5, r7 @ r5 = L_mpy_wx(cosx, t1) - L_mpy_wx(sinx, t0)@ 104 mov r11, r5 @ t1 = r5@ 106 sub r7, r7, r3 @ r5 [all...] |
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
H A D | idct.cpp | 131 int32 r0, r1, r2, r3, r4, r5, r6, r7, r8; /* butterfly nodes */ local 154 r5 = blk[B_SIZE * 7 + i]; 158 if (!(r1 | r2 | r3 | r4 | r5 | r6 | r7)) 182 r8 = W7 * (r4 + r5); 186 r5 = (r8 - (W1 + W7) * r5); 201 r6 = r5 + r7; 202 r5 -= r7; 209 r2 = (181 * (r4 + r5) + 128) >> 8; /* rounding */ 210 r4 = (181 * (r4 - r5) 353 int32 r0, r1, r2, r3, r4, r5, r6, r7, r8; /* butterfly nodes */ local [all...] |
/frameworks/rs/cpu_ref/ |
H A D | rsCpuIntrinsics_neon_YuvToRGB.S | 89 movw r5, #((16 * 149 + (128 >> 1) + 128 * 204) >> 1) 90 vdup.i16 q13, r5 91 movw r5, #((-16 * 149 + 128 * 50 + 128 * 104) >> 0) 92 vdup.i16 q14, r5 93 movw r5, #((16 * 149 + (128 << 2) + 128 * 254) >> 1) 94 vdup.i16 q15, r5 227 push {r4,r5} 228 ldr r5, [sp, #8] 233 add r0, r5, LSL #2 234 add r1, r5 [all...] |
/frameworks/av/media/libstagefright/codecs/aacenc/src/asm/ARMV7/ |
H A D | Radix4FFT_v7.s | 39 mov r5, r2, lsl #1 42 mov r5, r5, lsl #2 44 rsbeq r12, r5, r5, lsl #2 47 rsb r12, r5, r5, lsl #2 61 add r8, r8, r5 @ xptr += step@ 72 add r8, r8, r5 @ xptr += step@ 83 add r8, r8, r5 [all...] |