/external/clang/test/CodeGen/ |
H A D | arm-neon-fma.c | 10 // CHECK-LABEL: define <2 x float> @test_fma_order(<2 x float> %accum, <2 x float> %lhs, <2 x float> %rhs) #0 { 11 // CHECK: [[TMP0:%.*]] = bitcast <2 x float> %accum to <8 x i8> 19 float32x2_t test_fma_order(float32x2_t accum, float32x2_t lhs, float32x2_t rhs) { argument 20 return vfma_f32(accum, lhs, rhs); 23 // CHECK-LABEL: define <4 x float> @test_fmaq_order(<4 x float> %accum, <4 x float> %lhs, <4 x float> %rhs) #0 { 24 // CHECK: [[TMP0:%.*]] = bitcast <4 x float> %accum to <16 x i8> 32 float32x4_t test_fmaq_order(float32x4_t accum, float32x4_t lhs, float32x4_t rhs) { argument 33 return vfmaq_f32(accum, lhs, rhs);
|
H A D | mandel.c | 30 volatile double __complex__ accum; variable
|
H A D | vector.c | 32 __m64 accum = _mm_setzero_si64(); local 35 accum = _mm_add_pi32(p[i], accum); 37 __m64 accum2 = _mm_unpackhi_pi32(accum, accum); 38 accum = _mm_add_pi32(accum, accum2); 40 int result = _mm_cvtsi64_si32(accum);
|
/external/clang/test/Sema/ |
H A D | arm64-neon-args.c | 12 float32x4_t test_vmlsq_lane(float32x4_t accum, float32x4_t lhs, float32x2_t rhs) { argument 13 return vmlsq_lane_f32(accum, lhs, rhs, 1);
|
/external/eigen/unsupported/test/ |
H A D | cxx11_tensor_scan.cpp | 27 float accum = 0; local 30 VERIFY_IS_EQUAL(result(i), accum); local 31 accum += tensor(i); 33 accum += tensor(i); 34 VERIFY_IS_EQUAL(result(i), accum); local 38 accum = 1; 42 VERIFY_IS_EQUAL(result(i), accum); local 43 accum *= tensor(i); 45 accum *= tensor(i); 46 VERIFY_IS_EQUAL(result(i), accum); local 61 float accum = 0; local 64 VERIFY_IS_EQUAL(result(i, 1, 2, 3), accum); local 94 int accum = 0; local 97 VERIFY_IS_EQUAL(result(i), accum); local [all...] |
H A D | cxx11_tensor_reduction.cpp | 303 void reduce(const float val, float* accum) { *accum += val * val; } argument 305 float finalize(const float accum) const { return 1.0f / (accum + offset_); }
|
/external/tcpdump/ |
H A D | checksum.c | 41 accum = i << 2 43 accum <<= 1 44 if accum & 0x400: 45 accum ^= 0x633 46 crc_table.append(accum) 95 register uint16_t accum; local 100 accum = ((unsigned short) i << 2); 103 if ((accum <<= 1) & 0x400) accum ^= CRC10_POLYNOMIAL; 105 verify_crc10_table[i] = accum; 114 verify_crc10_cksum(uint16_t accum, const u_char *p, int length) argument [all...] |
/external/vboot_reference/futility/ |
H A D | cmd_pcr.c | 102 uint8_t accum[SHA256_DIGEST_SIZE * 2]; local 161 memcpy(accum, pcr, sizeof(pcr)); 162 parse_digest_or_die(accum + digest_size, digest_size, argv[i]); 165 print_digest(accum + digest_size, digest_size); 168 digest = DigestBuf(accum, digest_size * 2, digest_alg);
|
/external/skia/gm/ |
H A D | OverStroke.cpp | 90 float accum = 0.0f; local 96 while (accum < length) { 97 if (meas.getPosTan(accum, &pos, &tan)) { 104 accum += spacing;
|
/external/skia/src/core/ |
H A D | SkRect.cpp | 64 Sk4s min, max, accum; local 75 accum = max = min; 76 accum = accum * Sk4s(0); 81 accum = accum * xy; 89 * in which case we could eliminate accum entirely, and just check min and max for 92 if (is_finite(accum)) {
|
/external/webp/src/dsp/ |
H A D | rescaler.c | 38 int accum = wrk->x_add; local 43 wrk->frow[x_out] = right * wrk->x_add + (left - right) * accum; 46 accum -= wrk->x_sub; 47 if (accum < 0) { 52 accum += wrk->x_add; 55 assert(wrk->x_sub == 0 /* <- special case for src_width=1 */ || accum == 0); 69 int accum = 0; local 72 accum += wrk->x_add; 73 while (accum > 0) { 74 accum [all...] |
H A D | rescaler_mips32.c | 40 int accum, accum1; local 47 "li %[accum], 0 \n\t" 49 "addu %[accum], %[accum], %[x_add] \n\t" 51 "blez %[accum], 3f \n\t" 54 "subu %[accum], %[accum], %[x_sub] \n\t" 57 "bgtz %[accum], 2b \n\t" 59 "negu %[accum1], %[accum] \n\t" 70 : [accum]" 99 int accum; local [all...] |
/external/deqp/external/vulkancts/modules/vulkan/compute/ |
H A D | vktComputeBasicComputeShaderTests.cpp | 59 T accum = 1; local 61 accum *= v[i]; 62 return accum;
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
H A D | TensorFunctors.h | 101 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void reduce(const T t, T* accum) const { 103 *accum = sum_op(*accum, t); 106 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void reducePacket(const Packet& p, Packet* accum) const { 107 (*accum) = padd<Packet>(*accum, p); 118 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE T finalize(const T accum) const { 119 return accum; 149 EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void reduce(const T t, T* accum) { argument 151 *accum 155 reducePacket(const Packet& p, Packet* accum) argument [all...] |
H A D | TensorReductionCuda.h | 25 __device__ EIGEN_ALWAYS_INLINE void atomicReduce(T* output, T accum, R& reducer) { argument 31 reducer.reduce(accum, reinterpret_cast<T*>(&newval)); 39 reducer.reduce(accum, reinterpret_cast<T*>(&newval)); 48 reducer.reduce(accum, reinterpret_cast<T*>(&newval)); 56 reducer.reduce(accum, reinterpret_cast<T*>(&newval)); 84 __device__ inline void atomicReduce(half2* output, half2 accum, R<half>& reducer) { argument 87 reducer.reducePacket(accum, reinterpret_cast<half2*>(&newval)); 95 reducer.reducePacket(accum, reinterpret_cast<half2*>(&newval)); 104 __device__ inline void atomicReduce(float* output, float accum, SumReducer<float>&) { argument 106 atomicAdd(output, accum); 160 typename Self::CoeffReturnType accum = reducer.initialize(); local 171 reducer.reduce(__shfl_down(accum, offset, warpSize), &accum); local 236 half2 accum = reducer.template initializePacket<half2>(); local 247 reducer.reducePacket(__shfl_down(accum, offset, warpSize), &accum); local [all...] |
H A D | TensorScan.h | 227 typename Self::CoeffReturnType accum = self.accumulator().initialize(); local 232 data[curr] = self.accumulator().finalize(accum); 233 self.accumulator().reduce(self.inner().coeff(curr), &accum); 235 self.accumulator().reduce(self.inner().coeff(curr), &accum); 236 data[curr] = self.accumulator().finalize(accum); 258 typename Self::CoeffReturnType accum = self.accumulator().initialize(); local 262 data[curr] = self.accumulator().finalize(accum); 263 self.accumulator().reduce(self.inner().coeff(curr), &accum); 265 self.accumulator().reduce(self.inner().coeff(curr), &accum); 266 data[curr] = self.accumulator().finalize(accum); [all...] |
/external/glide/third_party/gif_encoder/src/main/java/com/bumptech/glide/gifencoder/ |
H A D | LZWEncoder.java | 110 byte[] accum = new byte[256]; field in class:LZWEncoder 123 accum[a_count++] = c; 232 outs.write(accum, 0, a_count);
|
/external/iproute2/netem/ |
H A D | maketable.c | 125 int accum=0; local 128 accum += *table; 129 *table++ = accum; 131 *total = accum;
|
/external/libvpx/libvpx/vp9/common/ |
H A D | vp9_thread_common.c | 312 void vp9_accumulate_frame_counts(FRAME_COUNTS *accum, argument 318 accum->y_mode[i][j] += counts->y_mode[i][j]; 322 accum->uv_mode[i][j] += counts->uv_mode[i][j]; 326 accum->partition[i][j] += counts->partition[i][j]; 335 accum->eob_branch[i][j][k][l][m] += 338 accum->coef[i][j][k][l][m][n] += counts->coef[i][j][k][l][m][n]; 346 accum->eob_branch[i][j][k][l][m] += 351 // accum->coef[i][j][k][l][m][n] += 357 accum->switchable_interp[i][j] += counts->switchable_interp[i][j]; 361 accum [all...] |
/external/mesa3d/src/gallium/state_trackers/wgl/ |
H A D | stw_pixelformat.c | 126 unsigned accum, 174 pfi->pfd.cAccumBits = 4*accum; 175 pfi->pfd.cAccumRedBits = accum; 176 pfi->pfd.cAccumGreenBits = accum; 177 pfi->pfd.cAccumBlueBits = accum; 178 pfi->pfd.cAccumAlphaBits = accum; 189 * since state trackers can allocate depth/stencil/accum buffers, we provide 199 pfi->stvis.accum_format = (accum) ? 215 * Add the depth/stencil/accum/ms variants for a particular color format. 121 stw_pixelformat_add( struct stw_device *stw_dev, boolean extended, const struct stw_pf_color_info *color, const struct stw_pf_depth_info *depth, unsigned accum, boolean doublebuffer, unsigned samples ) argument
|
/external/skia/src/opts/ |
H A D | SkBitmapFilter_opts.h | 24 const SkConvolutionFilter1D::ConvolutionFixed* filterValues, __m128i& accum, int r) { 34 accum = _mm_add_epi32(accum, t); 56 __m128i accum = _mm_setzero_si128(); local 79 accum = _mm_add_epi32(accum, t); 82 accum = _mm_add_epi32(accum, t); 97 accum = _mm_add_epi32(accum, 23 AccumRemainder(const unsigned char* pixelsLeft, const SkConvolutionFilter1D::ConvolutionFixed* filterValues, __m128i& accum, int r) argument [all...] |
/external/flac/libFLAC/ |
H A D | bitwriter.c | 75 uint32_t accum; /* accumulator; bits are right-justified; when full, accum is appended to buffer */ member in struct:FLAC__BitWriter 78 unsigned bits; /* # of used bits in accum */ 193 fprintf(out, "%01u", bw->accum & (1 << (bw->bits-j-1)) ? 1:0); 250 /* append bits as complete word to buffer, but don't change bw->accum or bw->bits */ 251 bw->buffer[bw->words] = SWAP_BE_WORD_TO_HOST(bw->accum << (FLAC__BITS_PER_WORD-bw->bits)); 282 bw->accum <<= n; 286 bw->buffer[bw->words++] = SWAP_BE_WORD_TO_HOST(bw->accum); 299 bw->accum = 0; 325 bw->accum << [all...] |
/external/libvpx/libvpx/vpx_dsp/arm/ |
H A D | avg_neon.c | 71 int32x4_t accum = vdupq_n_s32(0); local 76 accum = vabal_s16(accum, vget_low_s16(src0), zero); 77 accum = vabal_s16(accum, vget_high_s16(src0), zero); 78 accum = vabal_s16(accum, vget_low_s16(src8), zero); 79 accum = vabal_s16(accum, vget_high_s16(src8), zero); 86 const int64x2_t s0 = vpaddlq_s32(accum); // cascadin [all...] |
/external/libvpx/libvpx/vpx_dsp/x86/ |
H A D | avg_intrin_sse2.c | 291 __m128i accum = zero; local 300 accum = _mm_add_epi32(accum, sum); 304 { // cascading summation of accum 305 __m128i hi = _mm_srli_si128(accum, 8); 306 accum = _mm_add_epi32(accum, hi); 307 hi = _mm_srli_epi64(accum, 32); 308 accum = _mm_add_epi32(accum, h [all...] |
/external/mesa3d/src/mesa/swrast/ |
H A D | s_renderbuffer.c | 52 * This one multi-purpose function can allocate stencil, depth, accum, color 90 /* for accum buffer */ 419 _mesa_error(ctx, GL_OUT_OF_MEMORY, "Allocating accum buffer"); 486 GLboolean accum, 530 if (accum) { 482 _swrast_add_soft_renderbuffers(struct gl_framebuffer *fb, GLboolean color, GLboolean depth, GLboolean stencil, GLboolean accum, GLboolean alpha, GLboolean aux) argument
|