1/* 2 * Copyright (c) 2013 The WebM project authors. All Rights Reserved. 3 * 4 * Use of this source code is governed by a BSD-style license 5 * that can be found in the LICENSE file in the root of the source 6 * tree. An additional intellectual property rights grant can be found 7 * in the file PATENTS. All contributing project authors may 8 * be found in the AUTHORS file in the root of the source tree. 9 */ 10 11#include <stdlib.h> 12 13#include "./vpx_dsp_rtcd.h" 14#include "vpx/vpx_integer.h" 15#include "vpx_dsp/mips/common_dspr2.h" 16#include "vpx_dsp/mips/loopfilter_filters_dspr2.h" 17#include "vpx_dsp/mips/loopfilter_macros_dspr2.h" 18#include "vpx_dsp/mips/loopfilter_masks_dspr2.h" 19#include "vpx_mem/vpx_mem.h" 20 21#if HAVE_DSPR2 22void vpx_lpf_horizontal_8_dspr2(unsigned char *s, int pitch, 23 const uint8_t *blimit, const uint8_t *limit, 24 const uint8_t *thresh) { 25 uint32_t mask; 26 uint32_t hev, flat; 27 uint8_t i; 28 uint8_t *sp3, *sp2, *sp1, *sp0, *sq0, *sq1, *sq2, *sq3; 29 uint32_t thresh_vec, flimit_vec, limit_vec; 30 uint32_t uflimit, ulimit, uthresh; 31 uint32_t p1_f0, p0_f0, q0_f0, q1_f0; 32 uint32_t p3, p2, p1, p0, q0, q1, q2, q3; 33 uint32_t p0_l, p1_l, p2_l, p3_l, q0_l, q1_l, q2_l, q3_l; 34 uint32_t p0_r, p1_r, p2_r, p3_r, q0_r, q1_r, q2_r, q3_r; 35 36 uflimit = *blimit; 37 ulimit = *limit; 38 uthresh = *thresh; 39 40 /* create quad-byte */ 41 __asm__ __volatile__( 42 "replv.qb %[thresh_vec], %[uthresh] \n\t" 43 "replv.qb %[flimit_vec], %[uflimit] \n\t" 44 "replv.qb %[limit_vec], %[ulimit] \n\t" 45 46 : [thresh_vec] "=&r"(thresh_vec), [flimit_vec] "=&r"(flimit_vec), 47 [limit_vec] "=r"(limit_vec) 48 : [uthresh] "r"(uthresh), [uflimit] "r"(uflimit), [ulimit] "r"(ulimit)); 49 50 /* prefetch data for store */ 51 prefetch_store(s); 52 53 for (i = 0; i < 2; i++) { 54 sp3 = s - (pitch << 2); 55 sp2 = sp3 + pitch; 56 sp1 = sp2 + pitch; 57 sp0 = sp1 + pitch; 58 sq0 = s; 59 sq1 = s + pitch; 60 sq2 = sq1 + pitch; 61 sq3 = sq2 + pitch; 62 63 __asm__ __volatile__( 64 "lw %[p3], (%[sp3]) \n\t" 65 "lw %[p2], (%[sp2]) \n\t" 66 "lw %[p1], (%[sp1]) \n\t" 67 "lw %[p0], (%[sp0]) \n\t" 68 "lw %[q0], (%[sq0]) \n\t" 69 "lw %[q1], (%[sq1]) \n\t" 70 "lw %[q2], (%[sq2]) \n\t" 71 "lw %[q3], (%[sq3]) \n\t" 72 73 : [p3] "=&r"(p3), [p2] "=&r"(p2), [p1] "=&r"(p1), [p0] "=&r"(p0), 74 [q3] "=&r"(q3), [q2] "=&r"(q2), [q1] "=&r"(q1), [q0] "=&r"(q0) 75 : [sp3] "r"(sp3), [sp2] "r"(sp2), [sp1] "r"(sp1), [sp0] "r"(sp0), 76 [sq3] "r"(sq3), [sq2] "r"(sq2), [sq1] "r"(sq1), [sq0] "r"(sq0)); 77 78 filter_hev_mask_flatmask4_dspr2(limit_vec, flimit_vec, thresh_vec, p1, p0, 79 p3, p2, q0, q1, q2, q3, &hev, &mask, &flat); 80 81 if ((flat == 0) && (mask != 0)) { 82 filter1_dspr2(mask, hev, p1, p0, q0, q1, &p1_f0, &p0_f0, &q0_f0, &q1_f0); 83 84 __asm__ __volatile__( 85 "sw %[p1_f0], (%[sp1]) \n\t" 86 "sw %[p0_f0], (%[sp0]) \n\t" 87 "sw %[q0_f0], (%[sq0]) \n\t" 88 "sw %[q1_f0], (%[sq1]) \n\t" 89 90 : 91 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 92 [q1_f0] "r"(q1_f0), [sp1] "r"(sp1), [sp0] "r"(sp0), [sq0] "r"(sq0), 93 [sq1] "r"(sq1)); 94 } else if ((mask & flat) == 0xFFFFFFFF) { 95 /* left 2 element operation */ 96 PACK_LEFT_0TO3() 97 mbfilter_dspr2(&p3_l, &p2_l, &p1_l, &p0_l, &q0_l, &q1_l, &q2_l, &q3_l); 98 99 /* right 2 element operation */ 100 PACK_RIGHT_0TO3() 101 mbfilter_dspr2(&p3_r, &p2_r, &p1_r, &p0_r, &q0_r, &q1_r, &q2_r, &q3_r); 102 103 COMBINE_LEFT_RIGHT_0TO2() 104 105 __asm__ __volatile__( 106 "sw %[p2], (%[sp2]) \n\t" 107 "sw %[p1], (%[sp1]) \n\t" 108 "sw %[p0], (%[sp0]) \n\t" 109 "sw %[q0], (%[sq0]) \n\t" 110 "sw %[q1], (%[sq1]) \n\t" 111 "sw %[q2], (%[sq2]) \n\t" 112 113 : 114 : [p2] "r"(p2), [p1] "r"(p1), [p0] "r"(p0), [q0] "r"(q0), 115 [q1] "r"(q1), [q2] "r"(q2), [sp2] "r"(sp2), [sp1] "r"(sp1), 116 [sp0] "r"(sp0), [sq0] "r"(sq0), [sq1] "r"(sq1), [sq2] "r"(sq2)); 117 } else if ((flat != 0) && (mask != 0)) { 118 /* filtering */ 119 filter1_dspr2(mask, hev, p1, p0, q0, q1, &p1_f0, &p0_f0, &q0_f0, &q1_f0); 120 121 /* left 2 element operation */ 122 PACK_LEFT_0TO3() 123 mbfilter_dspr2(&p3_l, &p2_l, &p1_l, &p0_l, &q0_l, &q1_l, &q2_l, &q3_l); 124 125 /* right 2 element operation */ 126 PACK_RIGHT_0TO3() 127 mbfilter_dspr2(&p3_r, &p2_r, &p1_r, &p0_r, &q0_r, &q1_r, &q2_r, &q3_r); 128 129 if (mask & flat & 0x000000FF) { 130 __asm__ __volatile__( 131 "sb %[p2_r], (%[sp2]) \n\t" 132 "sb %[p1_r], (%[sp1]) \n\t" 133 "sb %[p0_r], (%[sp0]) \n\t" 134 "sb %[q0_r], (%[sq0]) \n\t" 135 "sb %[q1_r], (%[sq1]) \n\t" 136 "sb %[q2_r], (%[sq2]) \n\t" 137 138 : 139 : [p2_r] "r"(p2_r), [p1_r] "r"(p1_r), [p0_r] "r"(p0_r), 140 [q0_r] "r"(q0_r), [q1_r] "r"(q1_r), [q2_r] "r"(q2_r), 141 [sp2] "r"(sp2), [sp1] "r"(sp1), [sp0] "r"(sp0), [sq0] "r"(sq0), 142 [sq1] "r"(sq1), [sq2] "r"(sq2)); 143 } else if (mask & 0x000000FF) { 144 __asm__ __volatile__( 145 "sb %[p1_f0], (%[sp1]) \n\t" 146 "sb %[p0_f0], (%[sp0]) \n\t" 147 "sb %[q0_f0], (%[sq0]) \n\t" 148 "sb %[q1_f0], (%[sq1]) \n\t" 149 150 : 151 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 152 [q1_f0] "r"(q1_f0), [sp1] "r"(sp1), [sp0] "r"(sp0), 153 [sq0] "r"(sq0), [sq1] "r"(sq1)); 154 } 155 156 __asm__ __volatile__( 157 "srl %[p2_r], %[p2_r], 16 \n\t" 158 "srl %[p1_r], %[p1_r], 16 \n\t" 159 "srl %[p0_r], %[p0_r], 16 \n\t" 160 "srl %[q0_r], %[q0_r], 16 \n\t" 161 "srl %[q1_r], %[q1_r], 16 \n\t" 162 "srl %[q2_r], %[q2_r], 16 \n\t" 163 "srl %[p1_f0], %[p1_f0], 8 \n\t" 164 "srl %[p0_f0], %[p0_f0], 8 \n\t" 165 "srl %[q0_f0], %[q0_f0], 8 \n\t" 166 "srl %[q1_f0], %[q1_f0], 8 \n\t" 167 168 : [p2_r] "+r"(p2_r), [p1_r] "+r"(p1_r), [p0_r] "+r"(p0_r), 169 [q0_r] "+r"(q0_r), [q1_r] "+r"(q1_r), [q2_r] "+r"(q2_r), 170 [p1_f0] "+r"(p1_f0), [p0_f0] "+r"(p0_f0), [q0_f0] "+r"(q0_f0), 171 [q1_f0] "+r"(q1_f0) 172 :); 173 174 if (mask & flat & 0x0000FF00) { 175 __asm__ __volatile__( 176 "sb %[p2_r], +1(%[sp2]) \n\t" 177 "sb %[p1_r], +1(%[sp1]) \n\t" 178 "sb %[p0_r], +1(%[sp0]) \n\t" 179 "sb %[q0_r], +1(%[sq0]) \n\t" 180 "sb %[q1_r], +1(%[sq1]) \n\t" 181 "sb %[q2_r], +1(%[sq2]) \n\t" 182 183 : 184 : [p2_r] "r"(p2_r), [p1_r] "r"(p1_r), [p0_r] "r"(p0_r), 185 [q0_r] "r"(q0_r), [q1_r] "r"(q1_r), [q2_r] "r"(q2_r), 186 [sp2] "r"(sp2), [sp1] "r"(sp1), [sp0] "r"(sp0), [sq0] "r"(sq0), 187 [sq1] "r"(sq1), [sq2] "r"(sq2)); 188 } else if (mask & 0x0000FF00) { 189 __asm__ __volatile__( 190 "sb %[p1_f0], +1(%[sp1]) \n\t" 191 "sb %[p0_f0], +1(%[sp0]) \n\t" 192 "sb %[q0_f0], +1(%[sq0]) \n\t" 193 "sb %[q1_f0], +1(%[sq1]) \n\t" 194 195 : 196 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 197 [q1_f0] "r"(q1_f0), [sp1] "r"(sp1), [sp0] "r"(sp0), 198 [sq0] "r"(sq0), [sq1] "r"(sq1)); 199 } 200 201 __asm__ __volatile__( 202 "srl %[p1_f0], %[p1_f0], 8 \n\t" 203 "srl %[p0_f0], %[p0_f0], 8 \n\t" 204 "srl %[q0_f0], %[q0_f0], 8 \n\t" 205 "srl %[q1_f0], %[q1_f0], 8 \n\t" 206 207 : [p2] "+r"(p2), [p1] "+r"(p1), [p0] "+r"(p0), [q0] "+r"(q0), 208 [q1] "+r"(q1), [q2] "+r"(q2), [p1_f0] "+r"(p1_f0), 209 [p0_f0] "+r"(p0_f0), [q0_f0] "+r"(q0_f0), [q1_f0] "+r"(q1_f0) 210 :); 211 212 if (mask & flat & 0x00FF0000) { 213 __asm__ __volatile__( 214 "sb %[p2_l], +2(%[sp2]) \n\t" 215 "sb %[p1_l], +2(%[sp1]) \n\t" 216 "sb %[p0_l], +2(%[sp0]) \n\t" 217 "sb %[q0_l], +2(%[sq0]) \n\t" 218 "sb %[q1_l], +2(%[sq1]) \n\t" 219 "sb %[q2_l], +2(%[sq2]) \n\t" 220 221 : 222 : [p2_l] "r"(p2_l), [p1_l] "r"(p1_l), [p0_l] "r"(p0_l), 223 [q0_l] "r"(q0_l), [q1_l] "r"(q1_l), [q2_l] "r"(q2_l), 224 [sp2] "r"(sp2), [sp1] "r"(sp1), [sp0] "r"(sp0), [sq0] "r"(sq0), 225 [sq1] "r"(sq1), [sq2] "r"(sq2)); 226 } else if (mask & 0x00FF0000) { 227 __asm__ __volatile__( 228 "sb %[p1_f0], +2(%[sp1]) \n\t" 229 "sb %[p0_f0], +2(%[sp0]) \n\t" 230 "sb %[q0_f0], +2(%[sq0]) \n\t" 231 "sb %[q1_f0], +2(%[sq1]) \n\t" 232 233 : 234 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 235 [q1_f0] "r"(q1_f0), [sp1] "r"(sp1), [sp0] "r"(sp0), 236 [sq0] "r"(sq0), [sq1] "r"(sq1)); 237 } 238 239 __asm__ __volatile__( 240 "srl %[p2_l], %[p2_l], 16 \n\t" 241 "srl %[p1_l], %[p1_l], 16 \n\t" 242 "srl %[p0_l], %[p0_l], 16 \n\t" 243 "srl %[q0_l], %[q0_l], 16 \n\t" 244 "srl %[q1_l], %[q1_l], 16 \n\t" 245 "srl %[q2_l], %[q2_l], 16 \n\t" 246 "srl %[p1_f0], %[p1_f0], 8 \n\t" 247 "srl %[p0_f0], %[p0_f0], 8 \n\t" 248 "srl %[q0_f0], %[q0_f0], 8 \n\t" 249 "srl %[q1_f0], %[q1_f0], 8 \n\t" 250 251 : [p2_l] "+r"(p2_l), [p1_l] "+r"(p1_l), [p0_l] "+r"(p0_l), 252 [q0_l] "+r"(q0_l), [q1_l] "+r"(q1_l), [q2_l] "+r"(q2_l), 253 [p1_f0] "+r"(p1_f0), [p0_f0] "+r"(p0_f0), [q0_f0] "+r"(q0_f0), 254 [q1_f0] "+r"(q1_f0) 255 :); 256 257 if (mask & flat & 0xFF000000) { 258 __asm__ __volatile__( 259 "sb %[p2_l], +3(%[sp2]) \n\t" 260 "sb %[p1_l], +3(%[sp1]) \n\t" 261 "sb %[p0_l], +3(%[sp0]) \n\t" 262 "sb %[q0_l], +3(%[sq0]) \n\t" 263 "sb %[q1_l], +3(%[sq1]) \n\t" 264 "sb %[q2_l], +3(%[sq2]) \n\t" 265 266 : 267 : [p2_l] "r"(p2_l), [p1_l] "r"(p1_l), [p0_l] "r"(p0_l), 268 [q0_l] "r"(q0_l), [q1_l] "r"(q1_l), [q2_l] "r"(q2_l), 269 [sp2] "r"(sp2), [sp1] "r"(sp1), [sp0] "r"(sp0), [sq0] "r"(sq0), 270 [sq1] "r"(sq1), [sq2] "r"(sq2)); 271 } else if (mask & 0xFF000000) { 272 __asm__ __volatile__( 273 "sb %[p1_f0], +3(%[sp1]) \n\t" 274 "sb %[p0_f0], +3(%[sp0]) \n\t" 275 "sb %[q0_f0], +3(%[sq0]) \n\t" 276 "sb %[q1_f0], +3(%[sq1]) \n\t" 277 278 : 279 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 280 [q1_f0] "r"(q1_f0), [sp1] "r"(sp1), [sp0] "r"(sp0), 281 [sq0] "r"(sq0), [sq1] "r"(sq1)); 282 } 283 } 284 285 s = s + 4; 286 } 287} 288 289void vpx_lpf_vertical_8_dspr2(unsigned char *s, int pitch, 290 const uint8_t *blimit, const uint8_t *limit, 291 const uint8_t *thresh) { 292 uint8_t i; 293 uint32_t mask, hev, flat; 294 uint8_t *s1, *s2, *s3, *s4; 295 uint32_t prim1, prim2, sec3, sec4, prim3, prim4; 296 uint32_t thresh_vec, flimit_vec, limit_vec; 297 uint32_t uflimit, ulimit, uthresh; 298 uint32_t p3, p2, p1, p0, q3, q2, q1, q0; 299 uint32_t p1_f0, p0_f0, q0_f0, q1_f0; 300 uint32_t p0_l, p1_l, p2_l, p3_l, q0_l, q1_l, q2_l, q3_l; 301 uint32_t p0_r, p1_r, p2_r, p3_r, q0_r, q1_r, q2_r, q3_r; 302 303 uflimit = *blimit; 304 ulimit = *limit; 305 uthresh = *thresh; 306 307 /* create quad-byte */ 308 __asm__ __volatile__( 309 "replv.qb %[thresh_vec], %[uthresh] \n\t" 310 "replv.qb %[flimit_vec], %[uflimit] \n\t" 311 "replv.qb %[limit_vec], %[ulimit] \n\t" 312 313 : [thresh_vec] "=&r"(thresh_vec), [flimit_vec] "=&r"(flimit_vec), 314 [limit_vec] "=r"(limit_vec) 315 : [uthresh] "r"(uthresh), [uflimit] "r"(uflimit), [ulimit] "r"(ulimit)); 316 317 prefetch_store(s + pitch); 318 319 for (i = 0; i < 2; i++) { 320 s1 = s; 321 s2 = s + pitch; 322 s3 = s2 + pitch; 323 s4 = s3 + pitch; 324 s = s4 + pitch; 325 326 __asm__ __volatile__( 327 "lw %[p0], -4(%[s1]) \n\t" 328 "lw %[p1], -4(%[s2]) \n\t" 329 "lw %[p2], -4(%[s3]) \n\t" 330 "lw %[p3], -4(%[s4]) \n\t" 331 "lw %[q3], (%[s1]) \n\t" 332 "lw %[q2], (%[s2]) \n\t" 333 "lw %[q1], (%[s3]) \n\t" 334 "lw %[q0], (%[s4]) \n\t" 335 336 : [p3] "=&r"(p3), [p2] "=&r"(p2), [p1] "=&r"(p1), [p0] "=&r"(p0), 337 [q0] "=&r"(q0), [q1] "=&r"(q1), [q2] "=&r"(q2), [q3] "=&r"(q3) 338 : [s1] "r"(s1), [s2] "r"(s2), [s3] "r"(s3), [s4] "r"(s4)); 339 340 /* transpose p3, p2, p1, p0 341 original (when loaded from memory) 342 register -4 -3 -2 -1 343 p0 p0_0 p0_1 p0_2 p0_3 344 p1 p1_0 p1_1 p1_2 p1_3 345 p2 p2_0 p2_1 p2_2 p2_3 346 p3 p3_0 p3_1 p3_2 p3_3 347 348 after transpose 349 register 350 p0 p3_3 p2_3 p1_3 p0_3 351 p1 p3_2 p2_2 p1_2 p0_2 352 p2 p3_1 p2_1 p1_1 p0_1 353 p3 p3_0 p2_0 p1_0 p0_0 354 */ 355 __asm__ __volatile__( 356 "precrq.qb.ph %[prim1], %[p0], %[p1] \n\t" 357 "precr.qb.ph %[prim2], %[p0], %[p1] \n\t" 358 "precrq.qb.ph %[prim3], %[p2], %[p3] \n\t" 359 "precr.qb.ph %[prim4], %[p2], %[p3] \n\t" 360 361 "precrq.qb.ph %[p1], %[prim1], %[prim2] \n\t" 362 "precr.qb.ph %[p3], %[prim1], %[prim2] \n\t" 363 "precrq.qb.ph %[sec3], %[prim3], %[prim4] \n\t" 364 "precr.qb.ph %[sec4], %[prim3], %[prim4] \n\t" 365 366 "precrq.ph.w %[p0], %[p1], %[sec3] \n\t" 367 "precrq.ph.w %[p2], %[p3], %[sec4] \n\t" 368 "append %[p1], %[sec3], 16 \n\t" 369 "append %[p3], %[sec4], 16 \n\t" 370 371 : [prim1] "=&r"(prim1), [prim2] "=&r"(prim2), [prim3] "=&r"(prim3), 372 [prim4] "=&r"(prim4), [p0] "+r"(p0), [p1] "+r"(p1), [p2] "+r"(p2), 373 [p3] "+r"(p3), [sec3] "=&r"(sec3), [sec4] "=&r"(sec4) 374 :); 375 376 /* transpose q0, q1, q2, q3 377 original (when loaded from memory) 378 register +1 +2 +3 +4 379 q3 q3_0 q3_1 q3_2 q3_3 380 q2 q2_0 q2_1 q2_2 q2_3 381 q1 q1_0 q1_1 q1_2 q1_3 382 q0 q0_0 q0_1 q0_2 q0_3 383 384 after transpose 385 register 386 q3 q0_3 q1_3 q2_3 q3_3 387 q2 q0_2 q1_2 q2_2 q3_2 388 q1 q0_1 q1_1 q2_1 q3_1 389 q0 q0_0 q1_0 q2_0 q3_0 390 */ 391 __asm__ __volatile__( 392 "precrq.qb.ph %[prim1], %[q3], %[q2] \n\t" 393 "precr.qb.ph %[prim2], %[q3], %[q2] \n\t" 394 "precrq.qb.ph %[prim3], %[q1], %[q0] \n\t" 395 "precr.qb.ph %[prim4], %[q1], %[q0] \n\t" 396 397 "precrq.qb.ph %[q2], %[prim1], %[prim2] \n\t" 398 "precr.qb.ph %[q0], %[prim1], %[prim2] \n\t" 399 "precrq.qb.ph %[sec3], %[prim3], %[prim4] \n\t" 400 "precr.qb.ph %[sec4], %[prim3], %[prim4] \n\t" 401 402 "precrq.ph.w %[q3], %[q2], %[sec3] \n\t" 403 "precrq.ph.w %[q1], %[q0], %[sec4] \n\t" 404 "append %[q2], %[sec3], 16 \n\t" 405 "append %[q0], %[sec4], 16 \n\t" 406 407 : [prim1] "=&r"(prim1), [prim2] "=&r"(prim2), [prim3] "=&r"(prim3), 408 [prim4] "=&r"(prim4), [q3] "+r"(q3), [q2] "+r"(q2), [q1] "+r"(q1), 409 [q0] "+r"(q0), [sec3] "=&r"(sec3), [sec4] "=&r"(sec4) 410 :); 411 412 filter_hev_mask_flatmask4_dspr2(limit_vec, flimit_vec, thresh_vec, p1, p0, 413 p3, p2, q0, q1, q2, q3, &hev, &mask, &flat); 414 415 if ((flat == 0) && (mask != 0)) { 416 filter1_dspr2(mask, hev, p1, p0, q0, q1, &p1_f0, &p0_f0, &q0_f0, &q1_f0); 417 STORE_F0() 418 } else if ((mask & flat) == 0xFFFFFFFF) { 419 /* left 2 element operation */ 420 PACK_LEFT_0TO3() 421 mbfilter_dspr2(&p3_l, &p2_l, &p1_l, &p0_l, &q0_l, &q1_l, &q2_l, &q3_l); 422 423 /* right 2 element operation */ 424 PACK_RIGHT_0TO3() 425 mbfilter_dspr2(&p3_r, &p2_r, &p1_r, &p0_r, &q0_r, &q1_r, &q2_r, &q3_r); 426 427 STORE_F1() 428 } else if ((flat != 0) && (mask != 0)) { 429 filter1_dspr2(mask, hev, p1, p0, q0, q1, &p1_f0, &p0_f0, &q0_f0, &q1_f0); 430 431 /* left 2 element operation */ 432 PACK_LEFT_0TO3() 433 mbfilter_dspr2(&p3_l, &p2_l, &p1_l, &p0_l, &q0_l, &q1_l, &q2_l, &q3_l); 434 435 /* right 2 element operation */ 436 PACK_RIGHT_0TO3() 437 mbfilter_dspr2(&p3_r, &p2_r, &p1_r, &p0_r, &q0_r, &q1_r, &q2_r, &q3_r); 438 439 if (mask & flat & 0x000000FF) { 440 __asm__ __volatile__( 441 "sb %[p2_r], -3(%[s4]) \n\t" 442 "sb %[p1_r], -2(%[s4]) \n\t" 443 "sb %[p0_r], -1(%[s4]) \n\t" 444 "sb %[q0_r], (%[s4]) \n\t" 445 "sb %[q1_r], +1(%[s4]) \n\t" 446 "sb %[q2_r], +2(%[s4]) \n\t" 447 448 : 449 : [p2_r] "r"(p2_r), [p1_r] "r"(p1_r), [p0_r] "r"(p0_r), 450 [q0_r] "r"(q0_r), [q1_r] "r"(q1_r), [q2_r] "r"(q2_r), 451 [s4] "r"(s4)); 452 } else if (mask & 0x000000FF) { 453 __asm__ __volatile__( 454 "sb %[p1_f0], -2(%[s4]) \n\t" 455 "sb %[p0_f0], -1(%[s4]) \n\t" 456 "sb %[q0_f0], (%[s4]) \n\t" 457 "sb %[q1_f0], +1(%[s4]) \n\t" 458 459 : 460 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 461 [q1_f0] "r"(q1_f0), [s4] "r"(s4)); 462 } 463 464 __asm__ __volatile__( 465 "srl %[p2_r], %[p2_r], 16 \n\t" 466 "srl %[p1_r], %[p1_r], 16 \n\t" 467 "srl %[p0_r], %[p0_r], 16 \n\t" 468 "srl %[q0_r], %[q0_r], 16 \n\t" 469 "srl %[q1_r], %[q1_r], 16 \n\t" 470 "srl %[q2_r], %[q2_r], 16 \n\t" 471 "srl %[p1_f0], %[p1_f0], 8 \n\t" 472 "srl %[p0_f0], %[p0_f0], 8 \n\t" 473 "srl %[q0_f0], %[q0_f0], 8 \n\t" 474 "srl %[q1_f0], %[q1_f0], 8 \n\t" 475 476 : [p2_r] "+r"(p2_r), [p1_r] "+r"(p1_r), [p0_r] "+r"(p0_r), 477 [q0_r] "+r"(q0_r), [q1_r] "+r"(q1_r), [q2_r] "+r"(q2_r), 478 [p1_f0] "+r"(p1_f0), [p0_f0] "+r"(p0_f0), [q0_f0] "+r"(q0_f0), 479 [q1_f0] "+r"(q1_f0) 480 :); 481 482 if (mask & flat & 0x0000FF00) { 483 __asm__ __volatile__( 484 "sb %[p2_r], -3(%[s3]) \n\t" 485 "sb %[p1_r], -2(%[s3]) \n\t" 486 "sb %[p0_r], -1(%[s3]) \n\t" 487 "sb %[q0_r], (%[s3]) \n\t" 488 "sb %[q1_r], +1(%[s3]) \n\t" 489 "sb %[q2_r], +2(%[s3]) \n\t" 490 491 : 492 : [p2_r] "r"(p2_r), [p1_r] "r"(p1_r), [p0_r] "r"(p0_r), 493 [q0_r] "r"(q0_r), [q1_r] "r"(q1_r), [q2_r] "r"(q2_r), 494 [s3] "r"(s3)); 495 } else if (mask & 0x0000FF00) { 496 __asm__ __volatile__( 497 "sb %[p1_f0], -2(%[s3]) \n\t" 498 "sb %[p0_f0], -1(%[s3]) \n\t" 499 "sb %[q0_f0], (%[s3]) \n\t" 500 "sb %[q1_f0], +1(%[s3]) \n\t" 501 502 : 503 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 504 [q1_f0] "r"(q1_f0), [s3] "r"(s3)); 505 } 506 507 __asm__ __volatile__( 508 "srl %[p1_f0], %[p1_f0], 8 \n\t" 509 "srl %[p0_f0], %[p0_f0], 8 \n\t" 510 "srl %[q0_f0], %[q0_f0], 8 \n\t" 511 "srl %[q1_f0], %[q1_f0], 8 \n\t" 512 513 : [p2] "+r"(p2), [p1] "+r"(p1), [p0] "+r"(p0), [q0] "+r"(q0), 514 [q1] "+r"(q1), [q2] "+r"(q2), [p1_f0] "+r"(p1_f0), 515 [p0_f0] "+r"(p0_f0), [q0_f0] "+r"(q0_f0), [q1_f0] "+r"(q1_f0) 516 :); 517 518 if (mask & flat & 0x00FF0000) { 519 __asm__ __volatile__( 520 "sb %[p2_l], -3(%[s2]) \n\t" 521 "sb %[p1_l], -2(%[s2]) \n\t" 522 "sb %[p0_l], -1(%[s2]) \n\t" 523 "sb %[q0_l], (%[s2]) \n\t" 524 "sb %[q1_l], +1(%[s2]) \n\t" 525 "sb %[q2_l], +2(%[s2]) \n\t" 526 527 : 528 : [p2_l] "r"(p2_l), [p1_l] "r"(p1_l), [p0_l] "r"(p0_l), 529 [q0_l] "r"(q0_l), [q1_l] "r"(q1_l), [q2_l] "r"(q2_l), 530 [s2] "r"(s2)); 531 } else if (mask & 0x00FF0000) { 532 __asm__ __volatile__( 533 "sb %[p1_f0], -2(%[s2]) \n\t" 534 "sb %[p0_f0], -1(%[s2]) \n\t" 535 "sb %[q0_f0], (%[s2]) \n\t" 536 "sb %[q1_f0], +1(%[s2]) \n\t" 537 538 : 539 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 540 [q1_f0] "r"(q1_f0), [s2] "r"(s2)); 541 } 542 543 __asm__ __volatile__( 544 "srl %[p2_l], %[p2_l], 16 \n\t" 545 "srl %[p1_l], %[p1_l], 16 \n\t" 546 "srl %[p0_l], %[p0_l], 16 \n\t" 547 "srl %[q0_l], %[q0_l], 16 \n\t" 548 "srl %[q1_l], %[q1_l], 16 \n\t" 549 "srl %[q2_l], %[q2_l], 16 \n\t" 550 "srl %[p1_f0], %[p1_f0], 8 \n\t" 551 "srl %[p0_f0], %[p0_f0], 8 \n\t" 552 "srl %[q0_f0], %[q0_f0], 8 \n\t" 553 "srl %[q1_f0], %[q1_f0], 8 \n\t" 554 555 : [p2_l] "+r"(p2_l), [p1_l] "+r"(p1_l), [p0_l] "+r"(p0_l), 556 [q0_l] "+r"(q0_l), [q1_l] "+r"(q1_l), [q2_l] "+r"(q2_l), 557 [p1_f0] "+r"(p1_f0), [p0_f0] "+r"(p0_f0), [q0_f0] "+r"(q0_f0), 558 [q1_f0] "+r"(q1_f0) 559 :); 560 561 if (mask & flat & 0xFF000000) { 562 __asm__ __volatile__( 563 "sb %[p2_l], -3(%[s1]) \n\t" 564 "sb %[p1_l], -2(%[s1]) \n\t" 565 "sb %[p0_l], -1(%[s1]) \n\t" 566 "sb %[q0_l], (%[s1]) \n\t" 567 "sb %[q1_l], +1(%[s1]) \n\t" 568 "sb %[q2_l], +2(%[s1]) \n\t" 569 570 : 571 : [p2_l] "r"(p2_l), [p1_l] "r"(p1_l), [p0_l] "r"(p0_l), 572 [q0_l] "r"(q0_l), [q1_l] "r"(q1_l), [q2_l] "r"(q2_l), 573 [s1] "r"(s1)); 574 } else if (mask & 0xFF000000) { 575 __asm__ __volatile__( 576 "sb %[p1_f0], -2(%[s1]) \n\t" 577 "sb %[p0_f0], -1(%[s1]) \n\t" 578 "sb %[q0_f0], (%[s1]) \n\t" 579 "sb %[q1_f0], +1(%[s1]) \n\t" 580 581 : 582 : [p1_f0] "r"(p1_f0), [p0_f0] "r"(p0_f0), [q0_f0] "r"(q0_f0), 583 [q1_f0] "r"(q1_f0), [s1] "r"(s1)); 584 } 585 } 586 } 587} 588#endif // #if HAVE_DSPR2 589