Lines Matching refs:bs

48 static INLINE void d207_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
53 for (r = 0; r < bs - 1; ++r)
55 dst[(bs - 1) * stride] = left[bs - 1];
59 for (r = 0; r < bs - 2; ++r)
62 dst[(bs - 2) * stride] = ROUND_POWER_OF_TWO(left[bs - 2] +
63 left[bs - 1] * 3, 2);
64 dst[(bs - 1) * stride] = left[bs - 1];
68 for (c = 0; c < bs - 2; ++c)
69 dst[(bs - 1) * stride + c] = left[bs - 1];
71 for (r = bs - 2; r >= 0; --r)
72 for (c = 0; c < bs - 2; ++c)
77 static INLINE void d63_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
80 for (r = 0; r < bs; ++r) {
81 for (c = 0; c < bs; ++c)
92 static INLINE void d45_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
95 for (r = 0; r < bs; ++r) {
96 for (c = 0; c < bs; ++c)
97 dst[c] = r + c + 2 < bs * 2 ? ROUND_POWER_OF_TWO(above[r + c] +
100 : above[bs * 2 - 1];
106 static INLINE void d117_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
111 for (c = 0; c < bs; c++)
117 for (c = 1; c < bs; c++)
123 for (r = 3; r < bs; ++r)
128 for (r = 2; r < bs; ++r) {
129 for (c = 1; c < bs; c++)
136 static INLINE void d135_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
140 for (c = 1; c < bs; c++)
144 for (r = 2; r < bs; ++r)
149 for (r = 1; r < bs; ++r) {
150 for (c = 1; c < bs; c++)
157 static INLINE void d153_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
161 for (r = 1; r < bs; r++)
167 for (r = 2; r < bs; r++)
172 for (c = 0; c < bs - 2; c++)
176 for (r = 1; r < bs; ++r) {
177 for (c = 0; c < bs - 2; c++)
184 static INLINE void v_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
188 for (r = 0; r < bs; r++) {
189 vpx_memcpy(dst, above, bs);
195 static INLINE void h_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
199 for (r = 0; r < bs; r++) {
200 vpx_memset(dst, left[r], bs);
206 static INLINE void tm_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
211 for (r = 0; r < bs; r++) {
212 for (c = 0; c < bs; c++)
219 static INLINE void dc_128_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
223 for (r = 0; r < bs; r++) {
224 vpx_memset(dst, 128, bs);
230 static INLINE void dc_left_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
235 for (i = 0; i < bs; i++)
237 expected_dc = (sum + (bs >> 1)) / bs;
239 for (r = 0; r < bs; r++) {
240 vpx_memset(dst, expected_dc, bs);
246 static INLINE void dc_top_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
250 for (i = 0; i < bs; i++)
252 expected_dc = (sum + (bs >> 1)) / bs;
254 for (r = 0; r < bs; r++) {
255 vpx_memset(dst, expected_dc, bs);
261 static INLINE void dc_predictor(uint8_t *dst, ptrdiff_t stride, int bs,
264 const int count = 2 * bs;
266 for (i = 0; i < bs; i++) {
273 for (r = 0; r < bs; r++) {
274 vpx_memset(dst, expected_dc, bs);
323 const int bs = 4 << tx_size;
356 if (y0 + bs <= frame_height) {
357 for (i = 0; i < bs; ++i)
363 for (; i < bs; ++i)
368 for (i = 0; i < bs; ++i)
373 // TODO(hkuang) do not extend 2*bs pixels for all modes.
379 if (x0 + 2 * bs <= frame_width) {
380 if (right_available && bs == 4) {
381 vpx_memcpy(above_row, above_ref, 2 * bs);
383 vpx_memcpy(above_row, above_ref, bs);
384 vpx_memset(above_row + bs, above_row[bs - 1], bs);
386 } else if (x0 + bs <= frame_width) {
388 if (right_available && bs == 4) {
391 x0 + 2 * bs - frame_width);
393 vpx_memcpy(above_row, above_ref, bs);
394 vpx_memset(above_row + bs, above_row[bs - 1], bs);
398 if (right_available && bs == 4) {
401 x0 + 2 * bs - frame_width);
405 x0 + 2 * bs - frame_width);
411 if (bs == 4 && right_available && left_available) {
414 vpx_memcpy(above_row, above_ref, bs);
415 if (bs == 4 && right_available)
416 vpx_memcpy(above_row + bs, above_ref + bs, bs);
418 vpx_memset(above_row + bs, above_row[bs - 1], bs);
423 vpx_memset(above_row, 127, bs * 2);