/external/chromium_org/third_party/libwebp/dsp/ |
H A D | dec.c | 478 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 479 const int a = 3 * (q0 - p0) + VP8ksclip1[p1 - q1]; // in [-893,892] 483 p[ 0] = VP8kclip1[q0 - a1]; 488 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 489 const int a = 3 * (q0 - p0); 495 p[ 0] = VP8kclip1[q0 - a1]; 502 const int q0 = p[0], q1 = p[step], q2 = p[2*step]; local 503 const int a = VP8ksclip1[3 * (q0 - p0) + VP8ksclip1[p1 - q1]]; 511 p[ 0] = VP8kclip1[q0 - a1]; 517 const int p1 = p[-2*step], p0 = p[-step], q0 local 522 const int p1 = p[-2 * step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 529 const int p0 = p[-step], q0 = p[0]; local [all...] |
/external/chromium_org/third_party/mesa/src/src/mesa/tnl_dd/ |
H A D | t_dd_vertex.h | 65 GLfloat u0, v0, q0; member in struct:__anon14590::__anon14592
|
/external/fdlibm/ |
H A D | e_jn.c | 161 double q0,q1,h,tmp; int k,m; local 163 q0 = w; z = w+h; q1 = w*z - 1.0; k=1; 166 tmp = z*q1 - q0; 167 q0 = q1;
|
/external/libvpx/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copysrcframe_func_neon.asm | 53 vld1.8 {q0, q1}, [r2]! 63 vst1.8 {q0, q1}, [r3]! 110 vld1.8 {q0, q1}, [r2]! 116 vst1.8 {q0, q1}, [r3]! 167 vld1.8 {q0, q1}, [r2]! 173 vst1.8 {q0, q1}, [r3]! 216 vld1.8 {q0, q1}, [r2]! 220 vst1.8 {q0, q1}, [r3]!
|
/external/mesa3d/src/mesa/tnl_dd/ |
H A D | t_dd_vertex.h | 65 GLfloat u0, v0, q0; member in struct:__anon28001::__anon28003
|
/external/webp/src/dsp/ |
H A D | dec.c | 478 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 479 const int a = 3 * (q0 - p0) + VP8ksclip1[p1 - q1]; // in [-893,892] 483 p[ 0] = VP8kclip1[q0 - a1]; 488 const int p1 = p[-2*step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 489 const int a = 3 * (q0 - p0); 495 p[ 0] = VP8kclip1[q0 - a1]; 502 const int q0 = p[0], q1 = p[step], q2 = p[2*step]; local 503 const int a = VP8ksclip1[3 * (q0 - p0) + VP8ksclip1[p1 - q1]]; 511 p[ 0] = VP8kclip1[q0 - a1]; 517 const int p1 = p[-2*step], p0 = p[-step], q0 local 522 const int p1 = p[-2 * step], p0 = p[-step], q0 = p[0], q1 = p[step]; local 529 const int p0 = p[-step], q0 = p[0]; local [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/mips/dspr2/ |
H A D | vp9_mblpf_vert_loopfilter_dspr2.c | 34 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 87 "lw %[q0], (%[s4]) \n\t" 94 [q0] "=&r" (q0), [q7] "=&r" (q7), [q6] "=&r" (q6), 137 /* transpose q0, q1, q2, q3 143 q0 q0_0 q0_1 q0_2 q0_3 150 q0 q0_0 q1_0 q2_0 q3_0 155 "precrq.qb.ph %[prim3], %[q1], %[q0] \n\t" 156 "precr.qb.ph %[prim4], %[q1], %[q0] \n\t" 159 "precr.qb.ph %[q0], [all...] |
H A D | vp9_mblpf_horiz_loopfilter_dspr2.c | 36 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 98 "lw %[q0], (%[sq0]) \n\t" 107 : [q3] "=&r" (q3), [q2] "=&r" (q2), [q1] "=&r" (q1), [q0] "=&r" (q0), 114 p1, p0, p3, p2, q0, q1, q2, q3, 117 vp9_flatmask5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, &flat2); 122 vp9_filter1_dspr2(mask, hev, p1, p0, q0, q1, 180 "sw %[q0], (%[sq0]) \n\t" 184 [q2] "r" (q2), [q1] "r" (q1), [q0] "r" (q0), [all...] |
/external/chromium_org/third_party/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
H A D | filterbanks_neon.S | 52 vshll.s16 q0, d0, #16 57 vqadd.s32 q8, q1, q0 68 vshll.s16 q0, d0, #16 77 vqadd.s32 q8, q1, q0 84 vshll.s16 q0, d0, #16 90 vqadd.s32 q8, q1, q0 96 vshll.s16 q0, d0, #16 103 vqadd.s32 q8, q1, q0 106 vshll.s16 q0, d0, #16 114 vqadd.s32 q9, q1, q0 [all...] |
/external/libhevc/common/arm/ |
H A D | ihevc_inter_pred_chroma_copy_w16out.s | 134 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp) 137 vshl.i64 q0,q0,#6 @vshlq_n_s64(temp, 6) 175 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp) 178 vshl.i64 q0,q0,#6 @vshlq_n_s64(temp, 6) 217 vshl.i16 q0,q8,#6 @vshlq_n_s16(tmp, 6) 258 vshl.i16 q0,q8,#6 @vshlq_n_s16(tmp, 6) 292 vshl.i16 q0,q8,#6 @vshlq_n_s16(tmp, 6) 313 vshl.i16 q0,q [all...] |
H A D | ihevc_inter_pred_luma_copy_w16out.s | 104 vmovl.u8 q0,d0 @vmovl_u8(vld1_u8(pu1_src_tmp) 107 vshl.i64 q0,q0,#6 @vshlq_n_s64(temp, 6) 157 vshl.i16 q0,q8,#6 @vshlq_n_s16(tmp, 6) 198 vshl.i16 q0,q8,#6 @vshlq_n_s16(tmp, 6) 232 vshl.i16 q0,q8,#6 @vshlq_n_s16(tmp, 6)
|
H A D | ihevc_inter_pred_chroma_copy.s | 230 vld1.8 {q0},[r0]! @vld1_u8(pu1_src_tmp) 232 vst1.8 {q0},[r1]! @vst1_u8(pu1_dst_tmp, tmp_src) 257 vld1.8 {q0},[r0]! @vld1_u8(pu1_src_tmp) 259 vst1.8 {q0},[r1]! @vst1_u8(pu1_dst_tmp, tmp_src)
|
H A D | ihevc_inter_pred_chroma_vert_w16inp.s | 115 vmovl.s8 q0,d0 @long the value 139 vmull.s16 q0,d0,d12 @vmull_s16(src_tmp1, coeff_0) 144 vmlal.s16 q0,d2,d13 149 vmlal.s16 q0,d3,d14 151 vmlal.s16 q0,d6,d15 153 vqshrn.s32 d0,q0,#6 @right shift 155 vqrshrun.s16 d0,q0,#6 @rounding shift
|
H A D | ihevc_inter_pred_chroma_vert_w16inp_w16out.s | 115 vmovl.s8 q0,d0 @long the value 139 vmull.s16 q0,d0,d12 @vmull_s16(src_tmp1, coeff_0) 144 vmlal.s16 q0,d2,d13 149 vmlal.s16 q0,d3,d14 151 vmlal.s16 q0,d6,d15 153 vqshrn.s32 d0,q0,#6 @right shift
|
/external/libvpx/libvpx/vp9/common/mips/dspr2/ |
H A D | vp9_mblpf_vert_loopfilter_dspr2.c | 34 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 87 "lw %[q0], (%[s4]) \n\t" 94 [q0] "=&r" (q0), [q7] "=&r" (q7), [q6] "=&r" (q6), 137 /* transpose q0, q1, q2, q3 143 q0 q0_0 q0_1 q0_2 q0_3 150 q0 q0_0 q1_0 q2_0 q3_0 155 "precrq.qb.ph %[prim3], %[q1], %[q0] \n\t" 156 "precr.qb.ph %[prim4], %[q1], %[q0] \n\t" 159 "precr.qb.ph %[q0], [all...] |
H A D | vp9_mblpf_horiz_loopfilter_dspr2.c | 36 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 98 "lw %[q0], (%[sq0]) \n\t" 107 : [q3] "=&r" (q3), [q2] "=&r" (q2), [q1] "=&r" (q1), [q0] "=&r" (q0), 114 p1, p0, p3, p2, q0, q1, q2, q3, 117 vp9_flatmask5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, &flat2); 122 vp9_filter1_dspr2(mask, hev, p1, p0, q0, q1, 180 "sw %[q0], (%[sq0]) \n\t" 184 [q2] "r" (q2), [q1] "r" (q1), [q0] "r" (q0), [all...] |
/external/opencv/cv/src/ |
H A D | cvmoments.cpp | 539 double q0 = t0 * t0, q1 = t1 * t1; local 547 HuState->hu4 = q0 + q1; 548 HuState->hu6 = d * (q0 - q1) + n4 * t0 * t1; 550 t0 *= q0 - 3 * q1; 551 t1 *= 3 * q0 - q1; 553 q0 = nu30 - 3 * nu12; 556 HuState->hu3 = q0 * q0 + q1 * q1; 557 HuState->hu5 = q0 * t0 + q1 * t1; 558 HuState->hu7 = q1 * t0 - q0 * t [all...] |
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/arm/neon/ |
H A D | vp9_iht8x8_add_neon.asm | 120 ; will be stored back into q8-q15 registers. This macro will touch q0-q7 232 vadd.s16 q0, q9, q15 ; output[0] = step[0] + step[3] 271 vadd.s16 q8, q0, q7 ; output[0] = step1[0] + step1[7]; 278 vsub.s16 q15, q0, q7 ; output[7] = step1[0] - step1[7]; 284 ; q0 - q7 registers and use them as buffer during calculation. 380 vmull.s16 q0, d18, d30 384 vmlal.s16 q0, d28, d31 412 vadd.s32 q9, q4, q0 416 vsub.s32 q4, q4, q0 436 vmull.s16 q0, d [all...] |
H A D | vp9_loopfilter_neon.asm | 59 vld1.u8 {d7}, [r2@64], r1 ; q0 177 ; d7 q0 192 vabd.u8 d22, d16, d7 ; m4 = abs(q1 - q0) 200 vabd.u8 d17, d6, d7 ; abs(p0 - q0) 210 vcgt.u8 d22, d22, d2 ; (abs(q1 - q0) > thresh)*-1 214 vqadd.u8 d17, d17, d17 ; b = abs(p0 - q0) * 2 312 vld1.u8 {d7}, [r3@64], r1 ; q0 439 ; d7 q0 456 vabd.u8 d22, d16, d7 ; m4 = abs(q1 - q0) 468 vabd.u8 d26, d7, d17 ; m8 = abs(q0 [all...] |
/external/libvpx/libvpx/vp9/common/arm/neon/ |
H A D | vp9_iht8x8_add_neon.asm | 120 ; will be stored back into q8-q15 registers. This macro will touch q0-q7 232 vadd.s16 q0, q9, q15 ; output[0] = step[0] + step[3] 271 vadd.s16 q8, q0, q7 ; output[0] = step1[0] + step1[7]; 278 vsub.s16 q15, q0, q7 ; output[7] = step1[0] - step1[7]; 284 ; q0 - q7 registers and use them as buffer during calculation. 380 vmull.s16 q0, d18, d30 384 vmlal.s16 q0, d28, d31 412 vadd.s32 q9, q4, q0 416 vsub.s32 q4, q4, q0 436 vmull.s16 q0, d [all...] |
H A D | vp9_loopfilter_neon.asm | 59 vld1.u8 {d7}, [r2@64], r1 ; q0 177 ; d7 q0 192 vabd.u8 d22, d16, d7 ; m4 = abs(q1 - q0) 200 vabd.u8 d17, d6, d7 ; abs(p0 - q0) 210 vcgt.u8 d22, d22, d2 ; (abs(q1 - q0) > thresh)*-1 214 vqadd.u8 d17, d17, d17 ; b = abs(p0 - q0) * 2 312 vld1.u8 {d7}, [r3@64], r1 ; q0 439 ; d7 q0 456 vabd.u8 d22, d16, d7 ; m4 = abs(q1 - q0) 468 vabd.u8 d26, d7, d17 ; m8 = abs(q0 [all...] |
/external/chromium_org/third_party/opus/src/celt/arm/ |
H A D | celt_pitch_xcorr_arm.s | 49 ; q0 = opus_val32 sum[4] 51 ; q0 = opus_val32 sum[4] 86 VMLAL.S16 q0, d3, d6[0] 88 VMLAL.S16 q0, d4, d7[0] 90 VMLAL.S16 q0, d16, d6[1] 92 VMLAL.S16 q0, d17, d7[1] 94 VMLAL.S16 q0, d16, d6[2] 96 VMLAL.S16 q0, d17, d7[2] 98 VMLAL.S16 q0, d16, d6[3] 99 VMLAL.S16 q0, d1 [all...] |
/external/libopus/celt/arm/ |
H A D | celt_pitch_xcorr_arm.s | 49 ; q0 = opus_val32 sum[4] 51 ; q0 = opus_val32 sum[4] 86 VMLAL.S16 q0, d3, d6[0] 88 VMLAL.S16 q0, d4, d7[0] 90 VMLAL.S16 q0, d16, d6[1] 92 VMLAL.S16 q0, d17, d7[1] 94 VMLAL.S16 q0, d16, d6[2] 96 VMLAL.S16 q0, d17, d7[2] 98 VMLAL.S16 q0, d16, d6[3] 99 VMLAL.S16 q0, d1 [all...] |
/external/llvm/test/MC/AArch64/ |
H A D | inline-asm-modifiers.s | 132 ldr q0, [sp] 147 ldr q0, [sp]
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
H A D | variance_neon.asm | 36 vld1.8 {q0}, [r0], r1 ;Load up source and reference 68 vpaddl.s32 q0, q8 ;accumulate sum 109 vld1.8 {q0}, [r0], r1 ;Load up source and reference 138 vpaddl.s32 q0, q8 ;accumulate sum 193 vpaddl.s32 q0, q8 ;accumulate sum 258 vpaddl.s32 q0, q8 ;accumulate sum
|