/external/webp/src/dsp/ |
H A D | enc_mips32.c | 84 "lbu %[temp16], 0+" XSTR(BPS) "*" #A "(%[temp20]) \n\t" \ 85 "lbu %[temp17], 1+" XSTR(BPS) "*" #A "(%[temp20]) \n\t" \ 86 "lbu %[temp18], 2+" XSTR(BPS) "*" #A "(%[temp20]) \n\t" \ 87 "lbu %[temp19], 3+" XSTR(BPS) "*" #A "(%[temp20]) \n\t" \ 110 "sb %[" #TEMP0 "], 0+" XSTR(BPS) "*" #A "(%[temp16]) \n\t" \ 111 "sb %[" #TEMP4 "], 1+" XSTR(BPS) "*" #A "(%[temp16]) \n\t" \ 112 "sb %[" #TEMP8 "], 2+" XSTR(BPS) "*" #A "(%[temp16]) \n\t" \ 113 "sb %[" #TEMP12 "], 3+" XSTR(BPS) "*" #A "(%[temp16]) \n\t" 258 "lbu %[temp0], 0+" XSTR(BPS) "*" #A "(%[a]) \n\t" \ 259 "lbu %[temp1], 1+" XSTR(BPS) "*" # [all...] |
H A D | enc_mips_dsp_r2.c | 82 "lw %[" #TEMP2 "], " XSTR(BPS) "*" #A "(%[" #TEMP0 "]) \n\t" \ 83 "lw %[" #TEMP3 "], " XSTR(BPS) "*" #A "(%[" #TEMP1 "]) \n\t" \ 229 BPS) 234 dst, 0, 1, 2, 3, BPS) 259 BPS) 286 BPS) 322 for (y = 0; y < 16 * BPS; y += 4 * BPS) { 334 "usw %[value], 0+" #J "*" XSTR(BPS) "(%[dst]) \n\t" \ 335 "usw %[value], 4+" #J "*" XSTR(BPS) "( [all...] |
H A D | dec_mips32.c | 394 "lbu %[temp1], 0+0*" XSTR(BPS) "(%[dst]) \n\t" 402 "lbu %[temp18], 1+0*" XSTR(BPS) "(%[dst]) \n\t" 403 "sb %[temp1], 0+0*" XSTR(BPS) "(%[dst]) \n\t" 411 "lbu %[temp1], 2+0*" XSTR(BPS) "(%[dst]) \n\t" 412 "sb %[temp18], 1+0*" XSTR(BPS) "(%[dst]) \n\t" 420 "lbu %[temp18], 3+0*" XSTR(BPS) "(%[dst]) \n\t" 421 "sb %[temp1], 2+0*" XSTR(BPS) "(%[dst]) \n\t" 429 "sb %[temp18], 3+0*" XSTR(BPS) "(%[dst]) \n\t" 430 "lbu %[temp5], 0+1*" XSTR(BPS) "(%[dst]) \n\t" 431 "lbu %[temp8], 1+1*" XSTR(BPS) "( [all...] |
H A D | dec.c | 30 dst[(x) + (y) * BPS] = clip_8b(dst[(x) + (y) * BPS] + ((v) >> 3)) 79 dst += BPS; 109 VP8Transform(in + 2 * 16, dst + 4 * BPS, 1); 127 if (in[2 * 16]) VP8TransformDC(in + 2 * 16, dst + 4 * BPS); 128 if (in[3 * 16]) VP8TransformDC(in + 3 * 16, dst + 4 * BPS + 4); 170 #define DST(x, y) dst[(x) + (y) * BPS] 174 const uint8_t* top = dst - BPS; 183 dst += BPS; 196 memcpy(dst + j * BPS, ds [all...] |
H A D | dec_sse41.c | 29 dst += BPS;
|
H A D | enc.c | 36 0 + 0 * BPS, 4 + 0 * BPS, 8 + 0 * BPS, 12 + 0 * BPS, 37 0 + 4 * BPS, 4 + 4 * BPS, 8 + 4 * BPS, 12 + 4 * BPS, 38 0 + 8 * BPS, 4 + 8 * BPS, [all...] |
H A D | dec_sse2.c | 155 dst0 = _mm_loadl_epi64((__m128i*)(dst + 0 * BPS)); 156 dst1 = _mm_loadl_epi64((__m128i*)(dst + 1 * BPS)); 157 dst2 = _mm_loadl_epi64((__m128i*)(dst + 2 * BPS)); 158 dst3 = _mm_loadl_epi64((__m128i*)(dst + 3 * BPS)); 161 dst0 = _mm_cvtsi32_si128(WebPMemToUint32(dst + 0 * BPS)); 162 dst1 = _mm_cvtsi32_si128(WebPMemToUint32(dst + 1 * BPS)); 163 dst2 = _mm_cvtsi32_si128(WebPMemToUint32(dst + 2 * BPS)); 164 dst3 = _mm_cvtsi32_si128(WebPMemToUint32(dst + 3 * BPS)); 184 _mm_storel_epi64((__m128i*)(dst + 0 * BPS), dst0); 185 _mm_storel_epi64((__m128i*)(dst + 1 * BPS), dst [all...] |
H A D | enc_sse2.c | 153 ref0 = _mm_loadl_epi64((const __m128i*)&ref[0 * BPS]); 154 ref1 = _mm_loadl_epi64((const __m128i*)&ref[1 * BPS]); 155 ref2 = _mm_loadl_epi64((const __m128i*)&ref[2 * BPS]); 156 ref3 = _mm_loadl_epi64((const __m128i*)&ref[3 * BPS]); 159 ref0 = _mm_cvtsi32_si128(WebPMemToUint32(&ref[0 * BPS])); 160 ref1 = _mm_cvtsi32_si128(WebPMemToUint32(&ref[1 * BPS])); 161 ref2 = _mm_cvtsi32_si128(WebPMemToUint32(&ref[2 * BPS])); 162 ref3 = _mm_cvtsi32_si128(WebPMemToUint32(&ref[3 * BPS])); 182 _mm_storel_epi64((__m128i*)&dst[0 * BPS], ref0); 183 _mm_storel_epi64((__m128i*)&dst[1 * BPS], ref [all...] |
H A D | dec_neon.c | 393 vst1_lane_u32((uint32_t*)(dst + 0 * BPS), vreinterpret_u32_u8(dst01_u8), 0); 394 vst1_lane_u32((uint32_t*)(dst + 1 * BPS), vreinterpret_u32_u8(dst01_u8), 1); 395 vst1_lane_u32((uint32_t*)(dst + 2 * BPS), vreinterpret_u32_u8(dst23_u8), 0); 396 vst1_lane_u32((uint32_t*)(dst + 3 * BPS), vreinterpret_u32_u8(dst23_u8), 1); 406 dst01 = vld1_lane_u32((uint32_t*)(dst + 0 * BPS), dst01, 0); 407 dst23 = vld1_lane_u32((uint32_t*)(dst + 2 * BPS), dst23, 0); 408 dst01 = vld1_lane_u32((uint32_t*)(dst + 1 * BPS), dst01, 1); 409 dst23 = vld1_lane_u32((uint32_t*)(dst + 3 * BPS), dst23, 1); 1054 const int kBPS = BPS; 1283 const uint8x8_t A = vld1_u8(dst - BPS); // to [all...] |
H A D | dec_msa.c | 58 LD_SB4(dst, BPS, dest0, dest1, dest2, dest3); 67 ST4x4_UB(res0, res0, 3, 2, 1, 0, dst, BPS); 122 ADDBLK_ST4x4_UB(tmp0, tmp0, tmp0, tmp0, dst, BPS); 144 LD_UB4(dst, BPS, dest0, dest1, dest2, dest3); 153 ST4x4_UB(res0, res0, 3, 2, 1, 0, dst, BPS); 683 for (i = 0; i < 4; ++i) dc += dst[i - BPS] + dst[-1 + i * BPS]; 686 SW4(dc, dc, dc, dc, dst, BPS); 690 const uint8_t* const ptemp = dst - BPS - 1; 693 const v8i16 TL = (v8i16)__msa_fill_h(ptemp[0 * BPS]); [all...] |
H A D | enc_neon.c | 54 vst1_lane_u32((uint32_t*)(dst + 0 * BPS), vreinterpret_u32_u8(dst01_u8), 0); 55 vst1_lane_u32((uint32_t*)(dst + 1 * BPS), vreinterpret_u32_u8(dst01_u8), 1); 56 vst1_lane_u32((uint32_t*)(dst + 2 * BPS), vreinterpret_u32_u8(dst23_u8), 0); 57 vst1_lane_u32((uint32_t*)(dst + 3 * BPS), vreinterpret_u32_u8(dst23_u8), 1); 68 dst01 = vld1_lane_u32((uint32_t*)(ref + 0 * BPS), dst01, 0); 69 dst23 = vld1_lane_u32((uint32_t*)(ref + 2 * BPS), dst23, 0); 70 dst01 = vld1_lane_u32((uint32_t*)(ref + 1 * BPS), dst01, 1); 71 dst23 = vld1_lane_u32((uint32_t*)(ref + 3 * BPS), dst23, 1); 135 const int kBPS = BPS; 260 out = vld1q_lane_u32((const uint32_t*)(src + 0 * BPS), ou [all...] |
H A D | enc_msa.c | 60 LD_SB4(ref, BPS, dest0, dest1, dest2, dest3); 69 ST4x4_UB(res0, res0, 3, 2, 1, 0, dst, BPS); 94 LW4(src, BPS, in0, in1, in2, in3); 96 LW4(ref, BPS, in0, in1, in2, in3); 183 LW4(in, BPS, in0_m, in1_m, in2_m, in3_m); 214 for (y = 0; y < 16 * BPS; y += 4 * BPS) { 258 #define DST(x, y) dst[(x) + (y) * BPS] 272 SW4(out, out, out, out, dst, BPS); 281 WebPUint32ToMem(dst + 0 * BPS, [all...] |
H A D | enc_sse41.c | 80 const __m128i inA_0 = _mm_loadu_si128((const __m128i*)&inA[BPS * 0]); 81 const __m128i inA_1 = _mm_loadu_si128((const __m128i*)&inA[BPS * 1]); 82 const __m128i inA_2 = _mm_loadu_si128((const __m128i*)&inA[BPS * 2]); 87 const __m128i inA_3 = _mm_loadl_epi64((const __m128i*)&inA[BPS * 3]); 88 const __m128i inB_0 = _mm_loadu_si128((const __m128i*)&inB[BPS * 0]); 89 const __m128i inB_1 = _mm_loadu_si128((const __m128i*)&inB[BPS * 1]); 90 const __m128i inB_2 = _mm_loadu_si128((const __m128i*)&inB[BPS * 2]); 91 const __m128i inB_3 = _mm_loadl_epi64((const __m128i*)&inB[BPS * 3]); 181 for (y = 0; y < 16 * BPS; y += 4 * BPS) { [all...] |
H A D | dec_mips_dsp_r2.c | 33 BPS) 42 dst, 0, 1, 2, 3, BPS) 70 BPS) 77 temp7, temp6, dst, 0, 1, 2, 3, BPS) 140 BPS) 145 dst, 0, 1, 2, 3, BPS) 545 // TEMP0 = SRC[A + A1 * BPS] 546 // TEMP1 = SRC[B + B1 * BPS] 547 // TEMP2 = SRC[C + C1 * BPS] 548 // TEMP3 = SRC[D + D1 * BPS] [all...] |
H A D | dsp.h | 27 #define BPS 32 // this is the common stride for enc/dec macro 343 // *dst is the destination block, with stride BPS. Boundary samples are
|
/external/syslinux/diag/geodsp/ |
H A D | mk-lba-img.c | 26 #define BPS (512) macro 27 #define SECT_INT (BPS / sizeof(unsigned int)) 70 while ((len = fread(b, 1, BPS, stdin))) { 71 if (len < BPS) 72 memset((char *)b + len, 0, BPS - len); 73 fwrite(b, 1, BPS, f); 86 fwrite(b, 1, BPS, f);
|
/external/webp/src/enc/ |
H A D | filter_enc.c | 98 VP8SimpleHFilter16i(y_dst, BPS, limit); 99 VP8SimpleVFilter16i(y_dst, BPS, limit); 102 VP8HFilter16i(y_dst, BPS, limit, ilevel, hev_thresh); 103 VP8HFilter8i(u_dst, v_dst, BPS, limit, ilevel, hev_thresh); 104 VP8VFilter16i(y_dst, BPS, limit, ilevel, hev_thresh); 105 VP8VFilter8i(u_dst, v_dst, BPS, limit, ilevel, hev_thresh); 119 sum += VP8SSIMGetClipped(yuv1 + Y_OFF_ENC, BPS, yuv2 + Y_OFF_ENC, BPS, 125 sum += VP8SSIMGetClipped(yuv1 + U_OFF_ENC, BPS, yuv2 + U_OFF_ENC, BPS, [all...] |
H A D | vp8i_enc.h | 72 #define YUV_SIZE_ENC (BPS * 16) 73 #define PRED_SIZE_ENC (32 * BPS + 16 * BPS + 8 * BPS) // I16+Chroma+I4 preds 85 #define I16DC16 (0 * 16 * BPS) 87 #define I16VE16 (1 * 16 * BPS) 90 #define C8DC8 (2 * 16 * BPS) 92 #define C8VE8 (2 * 16 * BPS + 8 * BPS) 95 #define I4DC4 (3 * 16 * BPS [all...] |
H A D | quant_enc.c | 63 for (i = 0; i < 16; ++i) printf("%3d ", y_in[i + j * BPS]); 65 for (i = 0; i < 16; ++i) printf("%3d ", y_out[i + j * BPS]); 68 printf("%1d ", abs(y_in[i + j * BPS] - y_out[i + j * BPS])); 74 for (i = 0; i < 8; ++i) printf("%3d ", uv_in[i + j * BPS]); 76 for (i = 8; i < 16; ++i) printf("%3d ", uv_in[i + j * BPS]); 78 for (i = 0; i < 8; ++i) printf("%3d ", uv_out[i + j * BPS]); 80 for (i = 8; i < 16; ++i) printf("%3d ", uv_out[i + j * BPS]); 83 printf("%1d ", abs(uv_out[i + j * BPS] - uv_in[i + j * BPS])); [all...] |
H A D | iterator_enc.c | 108 dst += BPS; 112 memcpy(dst, dst - BPS, size); 113 dst += BPS; 177 src += BPS; 291 it->y_left_[i] = ysrc[15 + i * BPS]; 294 it->u_left_[i] = uvsrc[7 + i * BPS]; 295 it->v_left_[i] = uvsrc[15 + i * BPS]; 303 memcpy(it->y_top_, ysrc + 15 * BPS, 16); 304 memcpy(it->uv_top_, uvsrc + 7 * BPS, 8 + 8); 431 top[-4 + i] = blk[i + 3 * BPS]; // stor [all...] |
H A D | analysis_enc.c | 273 VP8Mean16x4(it->yuv_in_ + Y_OFF_ENC + k * BPS, &dc[k]);
|
/external/webp/src/dec/ |
H A D | frame_dec.c | 22 0 + 0 * BPS, 4 + 0 * BPS, 8 + 0 * BPS, 12 + 0 * BPS, 23 0 + 4 * BPS, 4 + 4 * BPS, 8 + 4 * BPS, 12 + 4 * BPS, 24 0 + 8 * BPS, 4 + 8 * BPS, [all...] |
H A D | vp8i_dec.h | 59 #define YUV_SIZE (BPS * 17 + BPS * 9) 60 #define Y_OFF (BPS * 1 + 8) 61 #define U_OFF (Y_OFF + BPS * 16 + BPS)
|
/external/llvm/lib/Transforms/Utils/ |
H A D | FlattenCFG.cpp | 211 BranchInst *BPS = dyn_cast<BranchInst>(PS->getTerminator()); local 212 if (BPS && BPS->isUnconditional()) {
|
H A D | Local.cpp | 1738 /// constructed in-place in the \c BPS map. Because of this \c BPS needs to 1743 /// Because we pass around references into \c BPS, we must use a container that 1748 std::map<Value *, Optional<BitPart>> &BPS) { 1749 auto I = BPS.find(V); 1750 if (I != BPS.end()) 1753 auto &Result = BPS[V] = None; 1760 MatchBitReversals, BPS); 1762 MatchBitReversals, BPS); 1795 MatchBitReversals, BPS); 1747 collectBitParts(Value *V, bool MatchBSwaps, bool MatchBitReversals, std::map<Value *, Optional<BitPart>> &BPS) argument 1905 std::map<Value *, Optional<BitPart>> BPS; local [all...] |