/external/clang/test/SemaCXX/ |
H A D | builtin-assume-aligned.cpp | 25 constexpr void *q4 = __builtin_assume_aligned(&n, 4, -4); variable
|
H A D | extern-c.cpp | 243 ExternCStruct4 *q4 = p4; // expected-error {{ambiguous}} member in namespace:tag_hiding
|
/external/capstone/suite/MC/ARM/ |
H A D | vfp4-thumb.s.cs | 5 0x08,0xef,0x50,0x4c = vfma.f32 q2, q4, q0 external variable declarations 11 0x28,0xef,0x50,0x4c = vfms.f32 q2, q4, q0 external variable declarations
|
H A D | vfp4.s.cs | 5 0x50,0x4c,0x08,0xf2 = vfma.f32 q2, q4, q0 external variable declarations 11 0x50,0x4c,0x28,0xf2 = vfms.f32 q2, q4, q0 external variable declarations
|
H A D | neon-v8.s.cs | 3 0x5c,0x4f,0x08,0xf3 = vmaxnm.f32 q2, q4, q6 external variable declarations 8 0x4c,0x80,0xbb,0xf3 = vcvta.s32.f32 q4, q6 9 0xe4,0x80,0xbb,0xf3 = vcvta.u32.f32 q4, q10 20 0x6e,0x82,0xbb,0xf3 = vcvtp.s32.f32 q4, q15 23 0x48,0x24,0xba,0xf3 = vrintn.f32 q1, q4 29 0xc8,0x25,0xfa,0xf3 = vrintz.f32 q9, q4 31 0xc8,0x26,0xba,0xf3 = vrintm.f32 q1, q4 33 0xc8,0x27,0xba,0xf3 = vrintp.f32 q1, q4 37 0xc8,0x25,0xfa,0xf3 = vrintz.f32 q9, q4 38 0xc8,0x27,0xba,0xf3 = vrintp.f32 q1, q4 [all...] |
H A D | thumb-neon-v8.s.cs | 3 0x08,0xff,0x5c,0x4f = vmaxnm.f32 q2, q4, q6 external variable declarations 8 0xbb,0xff,0x4c,0x80 = vcvta.s32.f32 q4, q6 9 0xbb,0xff,0xe4,0x80 = vcvta.u32.f32 q4, q10 20 0xbb,0xff,0x6e,0x82 = vcvtp.s32.f32 q4, q15 23 0xba,0xff,0x48,0x24 = vrintn.f32 q1, q4 29 0xfa,0xff,0xc8,0x25 = vrintz.f32 q9, q4 31 0xba,0xff,0xc8,0x26 = vrintm.f32 q1, q4 33 0xba,0xff,0xc8,0x27 = vrintp.f32 q1, q4 37 0xfa,0xff,0xc8,0x25 = vrintz.f32 q9, q4 38 0xba,0xff,0xc8,0x27 = vrintp.f32 q1, q4 [all...] |
H A D | neon-minmax-encoding.s.cs | 17 0x4c,0x86,0x1a,0xf2 = vmax.s16 q4, q5, q6 27 0x4a,0x86,0x18,0xf3 = vmax.u16 q4, q4, q5 external variable declarations 45 0x5c,0x86,0x1a,0xf2 = vmin.s16 q4, q5, q6 55 0x5a,0x86,0x18,0xf3 = vmin.u16 q4, q4, q5 external variable declarations
|
H A D | neon-mul-encoding.s.cs | 56 0x63,0x88,0xa8,0xf3 = vmul.i32 q4, q4, d3[1] external variable declarations 70 0x63,0xa8,0xa8,0xf3 = vmul.i32 q5, q4, d3[1] external variable declarations 71 0x44,0x88,0xaa,0xf3 = vmul.i32 q4, q5, d4[0]
|
H A D | neont2-minmax-encoding.s.cs | 17 0x1a,0xef,0x4c,0x86 = vmax.s16 q4, q5, q6 27 0x18,0xff,0x4a,0x86 = vmax.u16 q4, q4, q5 external variable declarations 45 0x1a,0xef,0x5c,0x86 = vmin.s16 q4, q5, q6 55 0x18,0xff,0x5a,0x86 = vmin.u16 q4, q4, q5 external variable declarations
|
H A D | neon-add-encoding.s.cs | 41 0x62,0x80,0x08,0xf3 = vhadd.u8 q4, q4, q9 external variable declarations
|
H A D | neon-sub-encoding.s.cs | 20 0x4e,0x88,0x38,0xf3 = vsub.i64 q4, q4, q7 external variable declarations 74 0x62,0x82,0x08,0xf3 = vhsub.u8 q4, q4, q9 external variable declarations
|
/external/flac/libFLAC/ |
H A D | lpc_intrin_avx2.c | 64 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 69 q4 = _mm256_set1_epi32(0xffff & qlp_coeff[4 ]); 87 mull = _mm256_madd_epi16(q4, _mm256_loadu_si256((const __m256i*)(data+i-5 ))); summ = _mm256_add_epi32(summ, mull); 97 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 102 q4 = _mm256_set1_epi32(0xffff & qlp_coeff[4 ]); 118 mull = _mm256_madd_epi16(q4, _mm256_loadu_si256((const __m256i*)(data+i-5 ))); summ = _mm256_add_epi32(summ, mull); 130 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 135 q4 = _mm256_set1_epi32(0xffff & qlp_coeff[4 ]); 149 mull = _mm256_madd_epi16(q4, _mm256_loadu_si256((const __m256i*)(data+i-5 ))); summ = _mm256_add_epi32(summ, mull); 159 __m256i q0, q1, q2, q3, q4, q local 190 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local 215 __m256i q0, q1, q2, q3, q4, q5, q6; local 240 __m256i q0, q1, q2, q3, q4, q5; local 261 __m256i q0, q1, q2, q3, q4; local 419 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 452 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 485 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 514 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 545 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local 570 __m256i q0, q1, q2, q3, q4, q5, q6; local 595 __m256i q0, q1, q2, q3, q4, q5; local 616 __m256i q0, q1, q2, q3, q4; local 778 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 811 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 844 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 873 __m256i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 904 __m256i q0, q1, q2, q3, q4, q5, q6, q7; local 929 __m256i q0, q1, q2, q3, q4, q5, q6; local 954 __m256i q0, q1, q2, q3, q4, q5; local 975 __m256i q0, q1, q2, q3, q4; local [all...] |
H A D | lpc_intrin_sse2.c | 70 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 75 q4 = _mm_cvtsi32_si128(0xffff & qlp_coeff[4]); q4 = _mm_shuffle_epi32(q4, _MM_SHUFFLE(0,0,0,0)); 93 mull = _mm_madd_epi16(q4, _mm_loadu_si128((const __m128i*)(data+i-5))); summ = _mm_add_epi32(summ, mull); 103 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 108 q4 = _mm_cvtsi32_si128(0xffff & qlp_coeff[4]); q4 = _mm_shuffle_epi32(q4, _MM_SHUFFLE(0,0,0,0)); 124 mull = _mm_madd_epi16(q4, _mm_loadu_si12 136 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 165 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 196 __m128i q0, q1, q2, q3, q4, q5, q6, q7; local 221 __m128i q0, q1, q2, q3, q4, q5, q6; local 246 __m128i q0, q1, q2, q3, q4, q5; local 267 __m128i q0, q1, q2, q3, q4; local [all...] |
H A D | lpc_intrin_sse41.c | 971 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 976 q4 = _mm_cvtsi32_si128(qlp_coeff[4]); q4 = _mm_shuffle_epi32(q4, _MM_SHUFFLE(0,0,0,0)); 994 mull = _mm_mullo_epi32(q4, _mm_loadu_si128((const __m128i*)(data+i-5))); summ = _mm_add_epi32(summ, mull); 1004 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10; local 1009 q4 = _mm_cvtsi32_si128(qlp_coeff[4]); q4 = _mm_shuffle_epi32(q4, _MM_SHUFFLE(0,0,0,0)); 1025 mull = _mm_mullo_epi32(q4, _mm_loadu_si12 1037 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8, q9; local 1066 __m128i q0, q1, q2, q3, q4, q5, q6, q7, q8; local 1097 __m128i q0, q1, q2, q3, q4, q5, q6, q7; local 1122 __m128i q0, q1, q2, q3, q4, q5, q6; local 1147 __m128i q0, q1, q2, q3, q4, q5; local 1168 __m128i q0, q1, q2, q3, q4; local [all...] |
/external/eigen/test/ |
H A D | geo_quaternion.cpp | 208 Quaternionx q4 = MCQuaternionUA(array3unaligned); local 212 VERIFY_IS_APPROX(q4.coeffs(), q3.coeffs());
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
H A D | dequant_idct_neon.c | 27 int16x8_t q1, q2, q3, q4, q5, q6; local 38 q4 = vld1q_s16(input); 59 vmulq_u16(vreinterpretq_u16_s16(q4), vreinterpretq_u16_s16(q6))); 67 q4 = vqdmulhq_n_s16(q2, cospi8sqrt2minus1); 69 q4 = vshrq_n_s16(q4, 1); 71 q4 = vqaddq_s16(q4, q2); 73 d10 = vqsub_s16(vget_low_s16(q3), vget_high_s16(q4)); 74 d11 = vqadd_s16(vget_high_s16(q3), vget_low_s16(q4)); [all...] |
H A D | idct_dequant_full_2x_neon.c | 21 int16x8_t q0, q1, q2, q3, q4, q5, q6, q7, q8, q9, q10, q11; local 41 q4 = vld1q_s16(q); 68 q4 = vmulq_s16(q4, q0); 74 dLow1 = vget_low_s16(q4); 75 dHigh1 = vget_high_s16(q4); 77 q4 = vcombine_s16(dHigh0, dHigh1); 86 q6 = vqdmulhq_n_s16(q4, sinpi8sqrt2); 88 q8 = vqdmulhq_n_s16(q4, cospi8sqrt2minus1); 97 q4 [all...] |
H A D | mbloopfilter_neon.c | 18 uint8x16_t q4, // p2 38 q11u8 = vabdq_u8(q3, q4); 39 q12u8 = vabdq_u8(q4, q5); 70 q4 = veorq_u8(q4, q0u8); 137 q0s8 = vqaddq_s8(vreinterpretq_s8_u8(q4), q0s8); 157 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 168 q4 = vld1q_u8(src); 182 vp8_mbloop_filter_neon(qblimit, qlimit, qthresh, q3, q4, q5, q6, q7, q8, q9, 183 q10, &q4, 14 vp8_mbloop_filter_neon(uint8x16_t qblimit, uint8x16_t qlimit, uint8x16_t qthresh, uint8x16_t q3, uint8x16_t q4, uint8x16_t q5, uint8x16_t q6, uint8x16_t q7, uint8x16_t q8, uint8x16_t q9, uint8x16_t q10, uint8x16_t *q4r, uint8x16_t *q5r, uint8x16_t *q6r, uint8x16_t *q7r, uint8x16_t *q8r, uint8x16_t *q9r) argument 205 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 292 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 456 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local [all...] |
H A D | vp8_loopfilter_neon.c | 19 uint8x16_t q4, // p2 36 q11u8 = vabdq_u8(q3, q4); 37 q12u8 = vabdq_u8(q4, q5); 41 q4 = vabdq_u8(q10, q9); 45 q3 = vmaxq_u8(q3, q4); 129 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 139 q4 = vld1q_u8(src); 153 vp8_loop_filter_neon(qblimit, qlimit, qthresh, q3, q4, q5, q6, q7, q8, q9, 172 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 216 q4 15 vp8_loop_filter_neon(uint8x16_t qblimit, uint8x16_t qlimit, uint8x16_t qthresh, uint8x16_t q3, uint8x16_t q4, uint8x16_t q5, uint8x16_t q6, uint8x16_t q7, uint8x16_t q8, uint8x16_t q9, uint8x16_t q10, uint8x16_t *q5r, uint8x16_t *q6r, uint8x16_t *q7r, uint8x16_t *q8r) argument 320 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local 432 uint8x16_t qblimit, qlimit, qthresh, q3, q4; local [all...] |
/external/libvpx/libvpx/vpx_dsp/mips/ |
H A D | loopfilter_masks_dspr2.h | 271 uint32_t q3, uint32_t q4, uint32_t *flat2) { 285 /* flat |= (abs(q4 - q0) > thresh) */ 286 "subu_s.qb %[c], %[q4], %[q0] \n\t" 287 "subu_s.qb %[r_k], %[q0], %[q4] \n\t" 345 [q0] "r"(q0), [q1] "r"(q1), [q2] "r"(q2), [q3] "r"(q3), [q4] "r"(q4), 269 flatmask5(uint32_t p4, uint32_t p3, uint32_t p2, uint32_t p1, uint32_t p0, uint32_t q0, uint32_t q1, uint32_t q2, uint32_t q3, uint32_t q4, uint32_t *flat2) argument
|
H A D | loopfilter_filters_dspr2.h | 514 const uint32_t q4 = *oq4, q5 = *oq5, q6 = *oq6, q7 = *oq7; local 522 /* addition of p6,p5,p4,p3,p2,p1,p0,q0,q1,q2,q3,q4,q5,q6 534 "addu.ph %[add_p6toq6], %[add_p6toq6], %[q4] \n\t" 542 [q3] "r"(q3), [q4] "r"(q4), [q5] "r"(q5), [q6] "r"(q6), 555 "subu.ph %[res_op6], %[res_op6], %[q4] \n\t" 569 "subu.ph %[res_op5], %[res_op5], %[q4] \n\t" 581 "subu.ph %[res_op4], %[res_op4], %[q4] \n\t" 591 "subu.ph %[res_op3], %[res_op3], %[q4] \n\t" 597 p0 + q0 + q1 + q2 + q3 + q4, [all...] |
H A D | loopfilter_mb_horiz_dspr2.c | 32 uint32_t p7, p6, p5, p4, p3, p2, p1, p0, q0, q1, q2, q3, q4, q5, q6, q7; local 96 "lw %[q4], (%[sq4]) \n\t" 102 [q7] "=&r"(q7), [q6] "=&r"(q6), [q5] "=&r"(q5), [q4] "=&r"(q4) 109 flatmask5(p7, p6, p5, p4, p0, q0, q4, q5, q6, q7, &flat2); 162 "sw %[q4], (%[sq4]) \n\t" 169 : [q6] "r"(q6), [q5] "r"(q5), [q4] "r"(q4), [q3] "r"(q3),
|
/external/guava/guava/src/com/google/common/cache/ |
H A D | Striped64.java | 100 volatile long q0, q1, q2, q3, q4, q5, q6; field in class:Striped64.Cell
|
/external/libjpeg-turbo/simd/ |
H A D | jsimd_arm_neon.S | 111 JLONG q1, q2, q3, q4, q5, q6, q7; \ 125 q4 = row5 + row1; \ define 127 MULTIPLY(q4, FIX_1_175875602); \ 129 MULTIPLY(q4, FIX_1_175875602_MINUS_0_390180644); \ 132 q4 = q6; \ define 138 q4 += MULTIPLY(row7, FIX_0_298631336_MINUS_0_899976223) + \ 165 tmp0 = q4; \ 279 vmov q4, q6 285 vmlsl.s16 q4, ROW1L, XFIX_0_899976223 320 vmlal.s16 q4, ROW7 [all...] |
/external/libvpx/libvpx/vpx_dsp/ |
H A D | loopfilter.c | 64 uint8_t q4) { 67 mask |= (abs(q4 - q0) > thresh) * -1; 243 const uint8_t q0 = *oq0, q1 = *oq1, q2 = *oq2, q3 = *oq3, q4 = *oq4, local 256 p7 * 3 + p6 + p5 + p4 + p3 + p2 * 2 + p1 + p0 + q0 + q1 + q2 + q3 + q4, 259 q0 + q1 + q2 + q3 + q4 + q5, 262 q1 + q2 + q3 + q4 + q5 + q6, 265 q2 + q3 + q4 + q5 + q6 + q7, 268 q3 + q4 + q5 + q6 + q7 * 2, 271 p4 + p3 + p2 + p1 + p0 + q0 + q1 + q2 * 2 + q3 + q4 + q5 + q6 + q7 * 3, 274 p3 + p2 + p1 + p0 + q0 + q1 + q2 + q3 * 2 + q4 61 flat_mask5(uint8_t thresh, uint8_t p4, uint8_t p3, uint8_t p2, uint8_t p1, uint8_t p0, uint8_t q0, uint8_t q1, uint8_t q2, uint8_t q3, uint8_t q4) argument 386 highbd_flat_mask5(uint8_t thresh, uint16_t p4, uint16_t p3, uint16_t p2, uint16_t p1, uint16_t p0, uint16_t q0, uint16_t q1, uint16_t q2, uint16_t q3, uint16_t q4, int bd) argument 596 const uint16_t q4 = *oq4; local [all...] |