Searched refs:s14 (Results 1 - 25 of 37) sorted by relevance

12

/external/compiler-rt/lib/builtins/arm/
H A Daddsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vadd.f32 s14, s14, s15
24 vmov r0, s14 // move result back to r0
H A Dsubsf3vfp.S22 vmov s14, r0 // move first param from r0 into float register
24 vsub.f32 s14, s14, s15
25 vmov r0, s14 // move result back to r0
H A Ddivsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vdiv.f32 s13, s14, s15
H A Deqsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
H A Dgesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
H A Dgtsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
H A Dlesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
H A Dltsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
H A Dmulsf3vfp.S21 vmov s14, r0 // move first param from r0 into float register
23 vmul.f32 s13, s14, s15
H A Dnesf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
H A Dunordsf2vfp.S22 vmov s14, r0 // move from GPR 0 to float register
24 vcmp.f32 s14, s15
/external/llvm/test/MC/AArch64/
H A Dneon-scalar-shift-imm.s73 sqshl s14, s17, #22
78 // CHECK: sqshl s14, s17, #22 // encoding: [0x2e,0x76,0x36,0x5f]
86 uqshl s14, s19, #18
91 // CHECK: uqshl s14, s19, #18 // encoding: [0x6e,0x76,0x32,0x7f]
99 sqshlu s16, s14, #25
104 // CHECK: sqshlu s16, s14, #25 // encoding: [0xd0,0x65,0x39,0x7f]
136 uqshrn h10, s14, #5
140 // CHECK: uqshrn h10, s14, #5 // encoding: [0xca,0x95,0x1b,0x7f]
169 sqshrun h20, s14, #3
173 // CHECK: sqshrun h20, s14, #
[all...]
H A Dneon-scalar-recip.s29 frecpe s19, s14
32 // CHECK: frecpe s19, s14 // encoding: [0xd3,0xd9,0xa1,0x5e]
H A Dneon-scalar-mul.s49 sqdmlsl s14, h12, h25
52 // CHECK: sqdmlsl s14, h12, h25 // encoding: [0x8e,0xb1,0x79,0x5e]
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/mips/dspr2/
H A Dvp9_itrans16_dspr2.c915 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; local
958 s14 = x14 * cospi_29_64 + x15 * cospi_3_64;
967 x6 = dct_const_round_shift(s6 + s14);
975 x14 = dct_const_round_shift(s6 - s14);
993 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64;
1006 x10 = dct_const_round_shift(s10 + s14);
1010 x14 = dct_const_round_shift(s10 - s14);
1028 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64;
1043 x12 = dct_const_round_shift(s12 + s14);
1045 x14 = dct_const_round_shift(s12 - s14);
[all...]
/external/libvpx/libvpx/vp9/common/mips/dspr2/
H A Dvp9_itrans16_dspr2.c915 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; local
958 s14 = x14 * cospi_29_64 + x15 * cospi_3_64;
967 x6 = dct_const_round_shift(s6 + s14);
975 x14 = dct_const_round_shift(s6 - s14);
993 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64;
1006 x10 = dct_const_round_shift(s10 + s14);
1010 x14 = dct_const_round_shift(s10 - s14);
1028 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64;
1043 x12 = dct_const_round_shift(s12 + s14);
1045 x14 = dct_const_round_shift(s12 - s14);
[all...]
/external/clang/test/CodeGen/
H A Dexprs.c150 void f14(struct s14 *a) {
/external/llvm/test/MC/ARM/
H A Dfp-armv8.s76 vselvs.f32 s21, s16, s14
77 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x07,0xaa,0x58,0xfe]
104 vrintxvs.f32 s10, s14
105 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0x47,0x5a,0xb7,0x6e]
H A Dthumb-fp-armv8.s79 vselvs.f32 s21, s16, s14
80 @ CHECK: vselvs.f32 s21, s16, s14 @ encoding: [0x58,0xfe,0x07,0xaa]
110 vrintxvs.f32 s10, s14
111 @ CHECK: vrintxvs.f32 s10, s14 @ encoding: [0xb7,0xee,0x47,0x5a]
/external/libvpx/libvpx/vp9/common/
H A Dvp9_idct.c639 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; local
682 s14 = x14 * cospi_29_64 + x15 * cospi_3_64;
691 x6 = dct_const_round_shift(s6 + s14);
699 x14 = dct_const_round_shift(s6 - s14);
717 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64;
730 x10 = dct_const_round_shift(s10 + s14);
734 x14 = dct_const_round_shift(s10 - s14);
752 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64;
767 x12 = dct_const_round_shift(s12 + s14);
769 x14 = dct_const_round_shift(s12 - s14);
[all...]
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/common/
H A Dvp9_idct.c675 tran_high_t s9, s10, s11, s12, s13, s14, s15; local
718 s14 = x14 * cospi_29_64 + x15 * cospi_3_64;
727 x6 = dct_const_round_shift(s6 + s14);
735 x14 = dct_const_round_shift(s6 - s14);
753 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64;
766 x10 = dct_const_round_shift(s10 + s14);
770 x14 = dct_const_round_shift(s10 - s14);
788 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64;
803 x12 = dct_const_round_shift(s12 + s14);
805 x14 = dct_const_round_shift(s12 - s14);
2094 tran_high_t s9, s10, s11, s12, s13, s14, s15; local
[all...]
/external/chromium_org/third_party/libvpx/source/libvpx/vp9/encoder/
H A Dvp9_dct.c840 tran_high_t s9, s10, s11, s12, s13, s14, s15; local
874 s14 = x14 * cospi_29_64 + x15 * cospi_3_64;
883 x6 = fdct_round_shift(s6 + s14);
891 x14 = fdct_round_shift(s6 - s14);
909 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64;
922 x10 = fdct_round_shift(s10 + s14);
926 x14 = fdct_round_shift(s10 - s14);
944 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64;
959 x12 = fdct_round_shift(s12 + s14);
961 x14 = fdct_round_shift(s12 - s14);
[all...]
/external/libvpx/libvpx/vp9/encoder/
H A Dvp9_dct.c799 int s0, s1, s2, s3, s4, s5, s6, s7, s8, s9, s10, s11, s12, s13, s14, s15; local
833 s14 = x14 * cospi_29_64 + x15 * cospi_3_64;
842 x6 = fdct_round_shift(s6 + s14);
850 x14 = fdct_round_shift(s6 - s14);
868 s14 = - x14 * cospi_12_64 + x15 * cospi_20_64;
881 x10 = fdct_round_shift(s10 + s14);
885 x14 = fdct_round_shift(s10 - s14);
903 s14 = - x14 * cospi_24_64 + x15 * cospi_8_64;
918 x12 = fdct_round_shift(s12 + s14);
920 x14 = fdct_round_shift(s12 - s14);
[all...]
/external/clang/test/Misc/
H A Ddiag-template-diffing.cpp429 template<typename T> struct s14 {}; struct
430 template<typename T> using a14 = s14<T>;
/external/chromium_org/third_party/libvpx/source/libvpx/vp8/common/x86/
H A Dloopfilter_block_sse2_x86_64.asm330 %define s14 [src + 2 * stride7]
390 movdqa xmm6, s14
773 movdqa s14, xmm5

Completed in 3836 milliseconds

12