/system/core/libpixelflinger/codeflinger/ |
H A D | load_store.cpp | 351 const int shift = (GGL_DITHER_BITS - (sbits-dbits)); local 352 if (shift>0) ADD(AL, 0, ireg, ireg, reg_imm(dither.reg, LSR, shift)); 353 else if (shift<0) ADD(AL, 0, ireg, ireg, reg_imm(dither.reg, LSL,-shift)); 359 int shift = sh-dbits; local 361 MOV(AL, 0, ireg, reg_imm(s.reg, LSR, shift)); 369 MOV(AL, 0, d.reg, reg_imm(s.reg, LSR, shift)); 371 ORR(AL, 0, d.reg, d.reg, reg_imm(s.reg, LSR, shift)); 375 int shift local [all...] |
H A D | ARMAssembler.cpp | 510 uint32_t ARMAssembler::reg_imm(int Rm, int type, uint32_t shift) argument 512 return ((shift&0x1F)<<7) | ((type&0x3)<<5) | (Rm&0xF); 546 uint32_t shift, int W) 550 reg_imm(abs(Rm), type, shift); 553 uint32_t ARMAssembler::reg_scale_post(int Rm, int type, uint32_t shift) argument 555 return (1<<25) | (((uint32_t(Rm)>>31)^1)<<23) | reg_imm(abs(Rm), type, shift); 545 reg_scale_pre(int Rm, int type, uint32_t shift, int W) argument
|
H A D | ARMAssemblerProxy.cpp | 93 uint32_t ARMAssemblerProxy::reg_imm(int Rm, int type, uint32_t shift) argument 95 return mTarget->reg_imm(Rm, type, shift); 122 uint32_t ARMAssemblerProxy::reg_scale_pre(int Rm, int type, uint32_t shift, int W) argument 124 return mTarget->reg_scale_pre(Rm, type, shift, W); 127 uint32_t ARMAssemblerProxy::reg_scale_post(int Rm, int type, uint32_t shift) argument 129 return mTarget->reg_scale_post(Rm, type, shift);
|
H A D | texturing.cpp | 515 const int shift = 31 - gglClz(tmu.format.size); local 541 MOV(LT, 0, width, imm(1 << shift)); 542 if (shift) 543 MOV(GE, 0, width, reg_imm(width, LSL, shift)); 552 // width = 1<<shift 562 MOV(GT, 0, width, imm(1 << shift)); 576 MOV(LT, 0, height, imm(1 << shift)); 577 if (shift) 578 MOV(GE, 0, height, reg_imm(height, LSL, shift)); 586 if (shift) { 773 int shift = 0; local 1202 int shift = incoming.l; local 1277 int shift = incoming.l; local 1315 int shift = incoming.l; local [all...] |
H A D | GGLAssembler.cpp | 569 const int shift = fragment.h <= 8 ? 0 : fragment.h-8; local 574 if (shift) { 576 reg_imm(mAlphaSource.reg, LSR, shift)); 583 if (shift) { 585 reg_imm(fragment.reg, LSR, shift)); 590 mAlphaSource.s -= shift; 703 const int shift = GGL_COLOR_BITS-fragment.size(); local 705 if (shift) CMP(AL, fragment.reg, reg_imm(ref, LSR, shift)); 918 int shift local [all...] |
H A D | Arm64Assembler.cpp | 374 uint32_t shift; local 380 shift = mAddrMode.reg_imm_type; 386 shift = 0; 398 case opADD: *mPC++ = A64_ADD_W(Rd, Rn, Rm, shift, amount); break; 399 case opAND: *mPC++ = A64_AND_W(Rd, Rn, Rm, shift, amount); break; 400 case opORR: *mPC++ = A64_ORR_W(Rd, Rn, Rm, shift, amount); break; 401 case opMVN: *mPC++ = A64_ORN_W(Rd, Rn, Rm, shift, amount); break; 402 case opSUB: *mPC++ = A64_SUB_W(Rd, Rn, Rm, shift, amount, s);break; 885 uint32_t ArmToArm64Assembler::reg_imm(int Rm, int type, uint32_t shift) argument 889 mAddrMode.reg_imm_shift = shift; 925 reg_scale_pre(int Rm, int type, uint32_t shift, int W) argument 1073 A64_ADD_X(uint32_t Rd, uint32_t Rn, uint32_t Rm, uint32_t shift, uint32_t amount) argument 1082 A64_ADD_IMM_X(uint32_t Rd, uint32_t Rn, uint32_t imm, uint32_t shift) argument 1089 A64_SUB_IMM_X(uint32_t Rd, uint32_t Rn, uint32_t imm, uint32_t shift) argument 1096 A64_ADD_W(uint32_t Rd, uint32_t Rn, uint32_t Rm, uint32_t shift, uint32_t amount) argument 1106 A64_SUB_W(uint32_t Rd, uint32_t Rn, uint32_t Rm, uint32_t shift, uint32_t amount, uint32_t setflag) argument 1127 A64_AND_W(uint32_t Rd, uint32_t Rn, uint32_t Rm, uint32_t shift, uint32_t amount) argument 1137 A64_ORR_W(uint32_t Rd, uint32_t Rn, uint32_t Rm, uint32_t shift, uint32_t amount) argument 1147 A64_ORN_W(uint32_t Rd, uint32_t Rn, uint32_t Rm, uint32_t shift, uint32_t amount) argument 1177 A64_MOVZ_X(uint32_t Rd, uint32_t imm, uint32_t shift) argument 1184 A64_MOVK_W(uint32_t Rd, uint32_t imm, uint32_t shift) argument 1191 A64_MOVZ_W(uint32_t Rd, uint32_t imm, uint32_t shift) argument [all...] |
H A D | MIPSAssembler.cpp | 234 uint32_t ArmToMipsAssembler::reg_imm(int Rm, int type, uint32_t shift) argument 238 amode.value = shift; 278 uint32_t shift, int W) 280 LOG_ALWAYS_FATAL_IF(W | type | shift, "reg_scale_pre adv modes not yet implemented"); 284 // amode.value = shift; 289 uint32_t ArmToMipsAssembler::reg_scale_post(int Rm, int type, uint32_t shift) argument 277 reg_scale_pre(int Rm, int type, uint32_t shift, int W) argument
|
/system/core/libpixelflinger/tests/gglmul/ |
H A D | gglmul_test.cpp | 104 int shift; member in struct:gglMulx_test_t 129 printf("Test x=0x%08x y=0x%08x shift=%2d :", 130 test->x, test->y, test->shift); 131 actual = gglMulx(test->x, test->y, test->shift); 133 ((int64_t)test->x * test->y + (1 << (test->shift-1))) >> test->shift; 146 int shift; member in struct:gglMulAddx_test_t 172 printf("Test x=0x%08x y=0x%08x shift=%2d a=0x%08x :", 173 test->x, test->y, test->shift, test->a); 174 actual = gglMulAddx(test->x, test->y,test->a, test->shift); 189 int shift; member in struct:gglMulSubx_test_t [all...] |
/system/core/include/private/pixelflinger/ |
H A D | ggl_fixed.h | 110 inline GGLfixed gglMulx(GGLfixed x, GGLfixed y, int shift) CONST; 111 inline GGLfixed gglMulx(GGLfixed x, GGLfixed y, int shift) { argument 113 if (__builtin_constant_p(shift)) { 118 : "%[x]"(x), [y]"r"(y), [lshift] "I"(32-shift), [rshift] "I"(shift) 126 : "%[x]"(x), [y]"r"(y), [lshift] "r"(32-shift), [rshift] "r"(shift) 133 inline GGLfixed gglMulAddx(GGLfixed x, GGLfixed y, GGLfixed a, int shift) CONST; 134 inline GGLfixed gglMulAddx(GGLfixed x, GGLfixed y, GGLfixed a, int shift) { argument 136 if (__builtin_constant_p(shift)) { 155 gglMulSubx(GGLfixed x, GGLfixed y, GGLfixed a, int shift) argument 197 gglMulx(GGLfixed a, GGLfixed b, int shift) argument 293 gglMulAddx(GGLfixed a, GGLfixed b, GGLfixed c, int shift) argument 365 gglMulSubx(GGLfixed a, GGLfixed b, GGLfixed c, int shift) argument 465 gglMulx(GGLfixed x, GGLfixed y, int shift) argument 482 gglMulAddx(GGLfixed x, GGLfixed y, GGLfixed a, int shift) argument 496 gglMulSubx(GGLfixed x, GGLfixed y, GGLfixed a, int shift) argument 526 gglMulx(GGLfixed a, GGLfixed b, int shift) argument 530 gglMulAddx(GGLfixed a, GGLfixed b, GGLfixed c, int shift) argument 534 gglMulSubx(GGLfixed a, GGLfixed b, GGLfixed c, int shift) argument [all...] |
/system/core/libpixelflinger/ |
H A D | buffer.cpp | 322 // and shift / construct the pixel 338 const int shift = (GGL_DITHER_BITS - (sbits-dbits)); local 339 if (shift >= 0) v += (dither >> shift) << sl; 340 else v += (dither << (-shift)) << sl; 368 // so we just have to shift the component to the right location 369 int shift = dh-sh; local 370 in |= v<<shift; 375 int shift = sh-dh; local 376 if (shift > [all...] |
H A D | fixed.cpp | 54 // shift right 1-bit to make room for the sign bit 62 int shift; local 63 x = gglRecipQNormalized(x, &shift); 64 shift += 16-q; 65 if (shift > 0) 66 x += 1L << (shift-1); // rounding 67 x >>= shift; local
|
H A D | scanline.cpp | 1080 void init(const context_t* c, int shift); 1087 void horz_clamp_iterator::init(const context_t* c, int shift) argument 1104 m_data += (tx.surface.stride*v) << shift;
|
H A D | trap.cpp | 696 const int shift = TRI_ITERATORS_BITS - TRI_FRACTION_BITS; local 701 edge->x = (x1 << shift) + (1LU << (TRI_ITERATORS_BITS-1)); 882 const int shift = FIXED_BITS - TRI_FRACTION_BITS; local 885 edge->x = x1 << shift; 1002 const int32_t shift = TRI_FRACTION_BITS + TRI_ITERATORS_BITS - FIXED_BITS; local 1006 GGLfixed l_min = gglMulAddx(left->x_incr, y - left->y_top, left->x, shift); 1013 GGLfixed r_min = gglMulAddx(right->x_incr, y - right->y_top, right->x, shift);
|