1/* NEON optimized code (C) COPYRIGHT 2009 Motorola 2 * 3 * Use of this source code is governed by a BSD-style license that can be 4 * found in the LICENSE file. 5 */ 6 7#include "SkBitmapProcState.h" 8#include "SkPerspIter.h" 9#include "SkShader.h" 10#include "SkUtilsArm.h" 11#include "SkBitmapProcState_utils.h" 12 13#include <arm_neon.h> 14 15extern const SkBitmapProcState::MatrixProc ClampX_ClampY_Procs_neon[]; 16extern const SkBitmapProcState::MatrixProc RepeatX_RepeatY_Procs_neon[]; 17 18static void decal_nofilter_scale_neon(uint32_t dst[], SkFixed fx, SkFixed dx, int count); 19static void decal_filter_scale_neon(uint32_t dst[], SkFixed fx, SkFixed dx, int count); 20 21// TILEX_PROCF(fx, max) SkClampMax((fx) >> 16, max) 22static inline int16x8_t sbpsm_clamp_tile8(int32x4_t low, int32x4_t high, unsigned max) { 23 int16x8_t res; 24 25 // get the hi 16s of all those 32s 26 res = vuzpq_s16(vreinterpretq_s16_s32(low), vreinterpretq_s16_s32(high)).val[1]; 27 28 // clamp 29 res = vmaxq_s16(res, vdupq_n_s16(0)); 30 res = vminq_s16(res, vdupq_n_s16(max)); 31 32 return res; 33} 34 35// TILEX_PROCF(fx, max) SkClampMax((fx) >> 16, max) 36static inline int32x4_t sbpsm_clamp_tile4(int32x4_t f, unsigned max) { 37 int32x4_t res; 38 39 // get the hi 16s of all those 32s 40 res = vshrq_n_s32(f, 16); 41 42 // clamp 43 res = vmaxq_s32(res, vdupq_n_s32(0)); 44 res = vminq_s32(res, vdupq_n_s32(max)); 45 46 return res; 47} 48 49// TILEY_LOW_BITS(fy, max) (((fy) >> 12) & 0xF) 50static inline int32x4_t sbpsm_clamp_tile4_low_bits(int32x4_t fx) { 51 int32x4_t ret; 52 53 ret = vshrq_n_s32(fx, 12); 54 55 /* We don't need the mask below because the caller will 56 * overwrite the non-masked bits 57 */ 58 //ret = vandq_s32(ret, vdupq_n_s32(0xF)); 59 60 return ret; 61} 62 63// TILEX_PROCF(fx, max) (((fx)&0xFFFF)*((max)+1)>> 16) 64static inline int16x8_t sbpsm_repeat_tile8(int32x4_t low, int32x4_t high, unsigned max) { 65 uint16x8_t res; 66 uint32x4_t tmpl, tmph; 67 68 // get the lower 16 bits 69 res = vuzpq_u16(vreinterpretq_u16_s32(low), vreinterpretq_u16_s32(high)).val[0]; 70 71 // bare multiplication, not SkFixedMul 72 tmpl = vmull_u16(vget_low_u16(res), vdup_n_u16(max+1)); 73 tmph = vmull_u16(vget_high_u16(res), vdup_n_u16(max+1)); 74 75 // extraction of the 16 upper bits 76 res = vuzpq_u16(vreinterpretq_u16_u32(tmpl), vreinterpretq_u16_u32(tmph)).val[1]; 77 78 return vreinterpretq_s16_u16(res); 79} 80 81// TILEX_PROCF(fx, max) (((fx)&0xFFFF)*((max)+1)>> 16) 82static inline int32x4_t sbpsm_repeat_tile4(int32x4_t f, unsigned max) { 83 uint16x4_t res; 84 uint32x4_t tmp; 85 86 // get the lower 16 bits 87 res = vmovn_u32(vreinterpretq_u32_s32(f)); 88 89 // bare multiplication, not SkFixedMul 90 tmp = vmull_u16(res, vdup_n_u16(max+1)); 91 92 // extraction of the 16 upper bits 93 tmp = vshrq_n_u32(tmp, 16); 94 95 return vreinterpretq_s32_u32(tmp); 96} 97 98// TILEX_LOW_BITS(fx, max) ((((fx) & 0xFFFF) * ((max) + 1) >> 12) & 0xF) 99static inline int32x4_t sbpsm_repeat_tile4_low_bits(int32x4_t fx, unsigned max) { 100 uint16x4_t res; 101 uint32x4_t tmp; 102 int32x4_t ret; 103 104 // get the lower 16 bits 105 res = vmovn_u32(vreinterpretq_u32_s32(fx)); 106 107 // bare multiplication, not SkFixedMul 108 tmp = vmull_u16(res, vdup_n_u16(max + 1)); 109 110 // shift and mask 111 ret = vshrq_n_s32(vreinterpretq_s32_u32(tmp), 12); 112 113 /* We don't need the mask below because the caller will 114 * overwrite the non-masked bits 115 */ 116 //ret = vandq_s32(ret, vdupq_n_s32(0xF)); 117 118 return ret; 119} 120 121#define MAKENAME(suffix) ClampX_ClampY ## suffix ## _neon 122#define TILEX_PROCF(fx, max) SkClampMax((fx) >> 16, max) 123#define TILEY_PROCF(fy, max) SkClampMax((fy) >> 16, max) 124#define TILEX_PROCF_NEON8(l, h, max) sbpsm_clamp_tile8(l, h, max) 125#define TILEY_PROCF_NEON8(l, h, max) sbpsm_clamp_tile8(l, h, max) 126#define TILEX_PROCF_NEON4(fx, max) sbpsm_clamp_tile4(fx, max) 127#define TILEY_PROCF_NEON4(fy, max) sbpsm_clamp_tile4(fy, max) 128#define TILEX_LOW_BITS(fx, max) (((fx) >> 12) & 0xF) 129#define TILEY_LOW_BITS(fy, max) (((fy) >> 12) & 0xF) 130#define TILEX_LOW_BITS_NEON4(fx, max) sbpsm_clamp_tile4_low_bits(fx) 131#define TILEY_LOW_BITS_NEON4(fy, max) sbpsm_clamp_tile4_low_bits(fy) 132#define CHECK_FOR_DECAL 133#include "SkBitmapProcState_matrix_neon.h" 134 135#define MAKENAME(suffix) RepeatX_RepeatY ## suffix ## _neon 136#define TILEX_PROCF(fx, max) SK_USHIFT16(((fx) & 0xFFFF) * ((max) + 1)) 137#define TILEY_PROCF(fy, max) SK_USHIFT16(((fy) & 0xFFFF) * ((max) + 1)) 138#define TILEX_PROCF_NEON8(l, h, max) sbpsm_repeat_tile8(l, h, max) 139#define TILEY_PROCF_NEON8(l, h, max) sbpsm_repeat_tile8(l, h, max) 140#define TILEX_PROCF_NEON4(fx, max) sbpsm_repeat_tile4(fx, max) 141#define TILEY_PROCF_NEON4(fy, max) sbpsm_repeat_tile4(fy, max) 142#define TILEX_LOW_BITS(fx, max) ((((fx) & 0xFFFF) * ((max) + 1) >> 12) & 0xF) 143#define TILEY_LOW_BITS(fy, max) ((((fy) & 0xFFFF) * ((max) + 1) >> 12) & 0xF) 144#define TILEX_LOW_BITS_NEON4(fx, max) sbpsm_repeat_tile4_low_bits(fx, max) 145#define TILEY_LOW_BITS_NEON4(fy, max) sbpsm_repeat_tile4_low_bits(fy, max) 146#include "SkBitmapProcState_matrix_neon.h" 147 148 149 150void decal_nofilter_scale_neon(uint32_t dst[], SkFixed fx, SkFixed dx, int count) { 151 if (count >= 8) { 152 // SkFixed is 16.16 fixed point 153 SkFixed dx8 = dx * 8; 154 int32x4_t vdx8 = vdupq_n_s32(dx8); 155 156 // setup lbase and hbase 157 int32x4_t lbase, hbase; 158 lbase = vdupq_n_s32(fx); 159 lbase = vsetq_lane_s32(fx + dx, lbase, 1); 160 lbase = vsetq_lane_s32(fx + dx + dx, lbase, 2); 161 lbase = vsetq_lane_s32(fx + dx + dx + dx, lbase, 3); 162 hbase = lbase + vdupq_n_s32(4 * dx); 163 164 do { 165 // store the upper 16 bits 166 vst1q_u32(dst, vreinterpretq_u32_s16( 167 vuzpq_s16(vreinterpretq_s16_s32(lbase), vreinterpretq_s16_s32(hbase)).val[1] 168 )); 169 170 // on to the next group of 8 171 lbase += vdx8; 172 hbase += vdx8; 173 dst += 4; // we did 8 elements but the result is twice smaller 174 count -= 8; 175 fx += dx8; 176 } while (count >= 8); 177 } 178 179 uint16_t* xx = (uint16_t*)dst; 180 for (int i = count; i > 0; --i) { 181 *xx++ = SkToU16(fx >> 16); fx += dx; 182 } 183} 184 185void decal_filter_scale_neon(uint32_t dst[], SkFixed fx, SkFixed dx, int count) { 186 if (count >= 8) { 187 SkFixed dx8 = dx * 8; 188 int32x4_t vdx8 = vdupq_n_s32(dx8); 189 190 int32x4_t wide_fx, wide_fx2; 191 wide_fx = vdupq_n_s32(fx); 192 wide_fx = vsetq_lane_s32(fx + dx, wide_fx, 1); 193 wide_fx = vsetq_lane_s32(fx + dx + dx, wide_fx, 2); 194 wide_fx = vsetq_lane_s32(fx + dx + dx + dx, wide_fx, 3); 195 196 wide_fx2 = vaddq_s32(wide_fx, vdupq_n_s32(4 * dx)); 197 198 while (count >= 8) { 199 int32x4_t wide_out; 200 int32x4_t wide_out2; 201 202 wide_out = vshlq_n_s32(vshrq_n_s32(wide_fx, 12), 14); 203 wide_out = wide_out | (vshrq_n_s32(wide_fx,16) + vdupq_n_s32(1)); 204 205 wide_out2 = vshlq_n_s32(vshrq_n_s32(wide_fx2, 12), 14); 206 wide_out2 = wide_out2 | (vshrq_n_s32(wide_fx2,16) + vdupq_n_s32(1)); 207 208 vst1q_u32(dst, vreinterpretq_u32_s32(wide_out)); 209 vst1q_u32(dst+4, vreinterpretq_u32_s32(wide_out2)); 210 211 dst += 8; 212 fx += dx8; 213 wide_fx += vdx8; 214 wide_fx2 += vdx8; 215 count -= 8; 216 } 217 } 218 219 if (count & 1) 220 { 221 SkASSERT((fx >> (16 + 14)) == 0); 222 *dst++ = (fx >> 12 << 14) | ((fx >> 16) + 1); 223 fx += dx; 224 } 225 while ((count -= 2) >= 0) 226 { 227 SkASSERT((fx >> (16 + 14)) == 0); 228 *dst++ = (fx >> 12 << 14) | ((fx >> 16) + 1); 229 fx += dx; 230 231 *dst++ = (fx >> 12 << 14) | ((fx >> 16) + 1); 232 fx += dx; 233 } 234} 235