/external/clang/test/PCH/ |
H A D | reinclude.cpp | 10 int q2 = A::y; variable
|
/external/llvm/test/MC/ARM/ |
H A D | neon-vswp.s | 4 vswp q1, q2 7 @ CHECK: vswp q1, q2 @ encoding: [0x44,0x20,0xb2,0xf3]
|
H A D | neon-minmax-encoding.s | 19 vmax.s8 q1, q2, q3 27 vmax.s8 q2, q3 30 vmax.u8 q11, q2 33 vmax.f32 q2, q1 49 @ CHECK: vmax.s8 q1, q2, q3 @ encoding: [0x46,0x26,0x04,0xf2] 56 @ CHECK: vmax.s8 q2, q2, q3 @ encoding: [0x46,0x46,0x04,0xf2] 59 @ CHECK: vmax.u8 q11, q11, q2 @ encoding: [0xc4,0x66,0x46,0xf3] 62 @ CHECK: vmax.f32 q2, q2, q [all...] |
H A D | neont2-minmax-encoding.s | 21 vmax.s8 q1, q2, q3 29 vmax.s8 q2, q3 32 vmax.u8 q11, q2 35 vmax.f32 q2, q1 51 @ CHECK: vmax.s8 q1, q2, q3 @ encoding: [0x04,0xef,0x46,0x26] 58 @ CHECK: vmax.s8 q2, q2, q3 @ encoding: [0x04,0xef,0x46,0x46] 61 @ CHECK: vmax.u8 q11, q11, q2 @ encoding: [0x46,0xff,0xc4,0x66] 64 @ CHECK: vmax.f32 q2, q2, q [all...] |
H A D | neon-shiftaccum-encoding.s | 7 vsra.s8 q7, q2, #8 16 vsra.u16 q2, q7, #6 25 vsra.s8 q2, #8 42 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x54,0xe1,0x88,0xf2] 51 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x5e,0x41,0x9a,0xf3] 59 @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x54,0x41,0x88,0xf2] 80 vrsra.s8 q1, q2, #8 81 vrsra.s16 q2, q3, #16 98 vrsra.s8 q2, # [all...] |
H A D | neont2-shiftaccum-encoding.s | 9 vsra.s8 q7, q2, #8 18 vsra.u16 q2, q7, #6 27 vsra.s8 q2, #8 44 @ CHECK: vsra.s8 q7, q2, #8 @ encoding: [0x88,0xef,0x54,0xe1] 53 @ CHECK: vsra.u16 q2, q7, #6 @ encoding: [0x9a,0xff,0x5e,0x41] 61 @ CHECK: vsra.s8 q2, q2, #8 @ encoding: [0x88,0xef,0x54,0x41] 83 vrsra.s8 q1, q2, #8 84 vrsra.s16 q2, q3, #16 101 vrsra.s8 q2, # [all...] |
H A D | vfp4.s | 18 @ ARM: vfma.f32 q2, q4, q0 @ encoding: [0x50,0x4c,0x08,0xf2] 19 @ THUMB: vfma.f32 q2, q4, q0 @ encoding: [0x08,0xef,0x50,0x4c] 20 vfma.f32 q2, q4, q0 44 @ ARM: vfms.f32 q2, q4, q0 @ encoding: [0x50,0x4c,0x28,0xf2] 45 @ THUMB: vfms.f32 q2, q4, q0 @ encoding: [0x28,0xef,0x50,0x4c] 46 vfms.f32 q2, q4, q0
|
/external/valgrind/main/none/tests/arm/ |
H A D | neon128.c | 357 TESTINSN_imm("vmov.i8 q2", q2, 0x7); 371 TESTINSN_imm("vmvn.i8 q2", q2, 0x7); 382 TESTINSN_imm("vorr.i16 q2", q2, 0x7); 412 TESTINSN_bin("vadd.i32 q0, q1, q2", q0, q1, i32, 24, q2, i32, 120); 413 TESTINSN_bin("vadd.i64 q0, q1, q2", q0, q1, i32, 140, q2, i3 [all...] |
H A D | neon128.stdout.exp | 6 vmov.i8 q2, #0x7 :: Qd 0x07070707 0x07070707 0x07070707 0x07070707 7 vmov.i8 q2, #0x7 :: Qd 0x07070707 0x07070707 0x07070707 0x07070707 31 vmvn.i8 q2, #0x7 :: Qd 0xf8f8f8f8 0xf8f8f8f8 0xf8f8f8f8 0xf8f8f8f8 32 vmvn.i8 q2, #0x7 :: Qd 0xf8f8f8f8 0xf8f8f8f8 0xf8f8f8f8 0xf8f8f8f8 50 vorr.i16 q2, #0x7 :: Qd 0x55575557 0x55575557 0x55575557 0x55575557 51 vorr.i16 q2, #0x7 :: Qd 0x151f191f 0x141f1f1f 0x131f1a1f 0x121f1e1f 95 vadd.i32 q0, q1, q2 :: Qd 0x00000090 0x00000090 0x00000090 0x00000090 Qm (i32)0x00000018 Qn (i32)0x00000078 96 vadd.i64 q0, q1, q2 :: Qd 0x00000104 0x00000104 0x00000104 0x00000104 Qm (i32)0x0000008c Qn (i32)0x00000078 97 vadd.i32 q0, q1, q2 :: Qd 0x00000104 0x00000104 0x00000104 0x00000104 Qm (i32)0x0000008c Qn (i32)0x00000078 98 vadd.i16 q0, q1, q2 [all...] |
/external/eigen/test/ |
H A D | geo_quaternion.cpp | 72 Quaternionx q1, q2; 73 q2.setIdentity(); 74 VERIFY_IS_APPROX(Quaternionx(Quaternionx::Identity()).coeffs(), q2.coeffs()); 76 VERIFY_IS_APPROX(q1.coeffs(), (q1*q2).coeffs()); 79 q1 *= q2; 82 q2 = AngleAxisx(a, v1.normalized()); 85 Scalar refangle = internal::abs(AngleAxisx(q1.inverse()*q2).angle()); 89 if((q1.coeffs()-q2.coeffs()).norm() > 10*largeEps) 91 VERIFY_IS_MUCH_SMALLER_THAN(internal::abs(q1.angularDistance(q2) - refangle), Scalar(1)); 96 VERIFY_IS_APPROX(q1 * q2 * v 189 Quaternionx q2 = MQuaternionA(array2); local 212 QuaternionUA *q2 = ::new(reinterpret_cast<void*>(array2)) QuaternionUA; local [all...] |
/external/libvpx/vp8/decoder/arm/neon/ |
H A D | dequant_idct_neon.asm | 41 vmul.i16 q2, q4, q6 45 vswp d3, d4 ;q2(vp[4] vp[12]) 47 vqdmulh.s16 q3, q2, d0[2] 48 vqdmulh.s16 q4, q2, d0[0] 56 vqadd.s16 q3, q3, q2 57 vqadd.s16 q4, q4, q2 76 vqdmulh.s16 q3, q2, d0[2] 77 vqdmulh.s16 q4, q2, d0[0] 87 vqadd.s16 q3, q3, q2 88 vqadd.s16 q4, q4, q2 [all...] |
H A D | idct_dequant_dc_full_2x_neon.asm | 28 vld1.16 {q2, q3}, [r0] ; l q 50 vmul.i16 q2, q2, q0 61 ; q2: l0r0 q3: l8r8 76 vqadd.s16 q10, q2, q3 ; a1 = 0 + 8 77 vqsub.s16 q11, q2, q3 ; b1 = 0 - 8 97 vqsub.s16 q2, q6, q5 105 vqadd.s16 q5, q11, q2 106 vqsub.s16 q6, q11, q2 129 vqadd.s16 q2, q [all...] |
/external/jmonkeyengine/engine/src/core/com/jme3/math/ |
H A D | Quaternion.java | 176 * @param q2
181 public Quaternion(Quaternion q1, Quaternion q2, float interp) {
argument 182 slerp(q1, q2, interp);
624 * @param q2
629 public Quaternion slerp(Quaternion q1, Quaternion q2, float t) {
argument 631 if (q1.x == q2.x && q1.y == q2.y && q1.z == q2.z && q1.w == q2.w) {
636 float result = (q1.x * q2 686 slerp(Quaternion q2, float changeAmnt) argument 736 nlerp(Quaternion q2, float blend) argument [all...] |
/external/libvpx/vp8/common/arm/neon/ |
H A D | loopfiltersimplehorizontaledge_neon.asm | 59 ;vqsub.s8 q2, q7, q6 ; ( qs0 - ps0) 60 vsubl.s8 q2, d14, d12 ; ( qs0 - ps0) 65 ;vmul.i8 q2, q2, q10 ; 3 * ( qs0 - ps0) 66 vadd.s16 q11, q2, q2 ; 3 * ( qs0 - ps0) 71 vadd.s16 q2, q2, q11 74 vaddw.s8 q2, q2, d [all...] |
H A D | shortidct4x4llm_neon.asm | 35 vld1.16 {q1, q2}, [r0] 38 vswp d3, d4 ;q2(vp[4] vp[12]) 40 vqdmulh.s16 q3, q2, d0[2] 41 vqdmulh.s16 q4, q2, d0[0] 49 vqadd.s16 q3, q3, q2 ;modify since sinpi8sqrt2 > 65536/2 (negtive number) 50 vqadd.s16 q4, q4, q2 72 vqdmulh.s16 q3, q2, d0[2] 73 vqdmulh.s16 q4, q2, d0[0] 81 vqadd.s16 q3, q3, q2 ;modify since sinpi8sqrt2 > 65536/2 (negtive number) 82 vqadd.s16 q4, q4, q2 [all...] |
H A D | loopfiltersimpleverticaledge_neon.asm | 79 ;vqsub.s8 q2, q5, q4 ; ( qs0 - ps0) 80 vsubl.s8 q2, d8, d10 ; ( qs0 - ps0) 85 ;vmul.i8 q2, q2, q11 ; vp8_filter = vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0)) 86 vadd.s16 q10, q2, q2 ; 3 * ( qs0 - ps0) 88 vadd.s16 q2, q2, q10 91 ;vqadd.s8 q1, q1, q2 92 vaddw.s8 q2, q [all...] |
H A D | buildintrapredictorsmby_neon.asm | 60 vpaddl.u8 q2, q1 61 vpaddl.u16 q3, q2 180 vdup.u8 q2, r5 184 vst1.u8 {q2}, [r1]! 193 vdup.u8 q2, r5 197 vst1.u8 {q2}, [r1]! 207 vdup.u8 q2, r5 211 vst1.u8 {q2}, [r1]! 220 vdup.u8 q2, r5 224 vst1.u8 {q2}, [r [all...] |
H A D | recon16x16mb_neon.asm | 37 vmovl.u8 q2, d26 52 vadd.s16 q2, q2, q10 61 vqmovun.s16 d2, q2 69 vst1.u8 {q2}, [r2], r3 88 vmovl.u8 q2, d26 99 vadd.s16 q2, q2, q10 108 vqmovun.s16 d2, q2
|
H A D | loopfilter_neon.asm | 49 vld1.u8 {q9}, [r2], r1 ; q2 88 vld1.u8 {d18}, [r3], r1 ; q2 100 vld1.u8 {d19}, [r12], r1 ; q2 301 ; q2 thresh 308 ; q9 q2 318 vabd.u8 q3, q9, q8 ; abs(q2 - q1) 319 vabd.u8 q4, q10, q9 ; abs(q3 - q2) 328 vcgt.u8 q13, q13, q2 ; (abs(p1 - p0) > thresh)*-1 329 vcgt.u8 q14, q14, q2 ; (abs(q1 - q0) > thresh)*-1 336 vabd.u8 q2, q [all...] |
/external/libvpx/vpx_scale/arm/neon/ |
H A D | vp8_vpxyv12_copyframeyonly_neon.asm | 50 vld1.8 {q2, q3}, [r8]! 61 vst1.8 {q2, q3}, [r9]! 120 vmov q3, q2 131 vst1.8 {q2, q3}, [r6], lr 155 vld1.8 {q2, q3}, [r1]! 169 vst1.8 {q2, q3}, [r5]! 201 vld1.8 {q2}, [r2]! 209 vst1.8 {q2}, [r6], lr 211 vst1.8 {q2}, [r6], lr 213 vst1.8 {q2}, [r [all...] |
/external/clang/test/CodeGenCXX/ |
H A D | copy-assign-synthesis.cpp | 62 q1.pr(); q2.pr(); 65 Q q1, q2; member in struct:X
|
H A D | copy-assign-synthesis-1.cpp | 78 q1.pr(); q2.pr(); 81 Q q1, q2; member in struct:X
|
/external/libvpx/vp8/encoder/arm/neon/ |
H A D | vp8_subpixelvariance16x16s_neon.asm | 56 vext.8 q3, q2, q3, #1 62 vrhadd.u8 q1, q2, q3 63 vrhadd.u8 q2, q4, q5 72 vsubl.u8 q2, d6, d28 97 vpadal.s16 q8, q2 145 vld1.u8 {q2}, [r0], r1 153 vrhadd.u8 q0, q0, q2 155 vrhadd.u8 q2, q2, q4 165 vsubl.u8 q2, d1 [all...] |
/external/webkit/Source/WebCore/platform/graphics/gpu/ |
H A D | LoopBlinnMathUtils.h | 44 // that between (p2, q2). 48 const FloatPoint& q2);
|
/external/llvm/include/llvm/Analysis/ |
H A D | Trace.h | 105 iterator erase(iterator q1, iterator q2) { return BasicBlocks.erase (q1, q2); } argument
|