Lines Matching refs:srcp

233 	Uint32 *srcp = (Uint32 *)info->s_pixels;
249 Uint32 s = *srcp++;
257 movq_m2r((*srcp), mm1);/* 2 x src -> mm1(ARGBARGB) */
271 srcp += 2;
273 srcp += srcskip;
291 Uint32 *srcp = (Uint32 *)info->s_pixels;
312 movd_m2r((*srcp), mm1);/* src(ARGB) -> mm1 (0000ARGB)*/
325 ++srcp;
329 movq_m2r((*srcp), mm0);/* 2 x src -> mm0(ARGBARGB)*/
354 srcp += 2;
357 srcp += srcskip;
369 Uint32 *srcp = (Uint32 *)info->s_pixels;
395 Uint32 alpha = *srcp & amask;
405 movd_m2r((*srcp), mm1);/* src(ARGB) -> mm1 (0000ARGB)*/
412 movd_m2r((*srcp), mm1);/* src(ARGB) -> mm1 (0000ARGB)*/
435 ++srcp;
438 srcp += srcskip;
451 Uint32 *srcp = (Uint32 *)info->s_pixels;
466 Uint32 s = *srcp++;
477 src1 = *(__m64*)srcp; /* 2 x src -> src1(ARGBARGB) */
492 srcp += 2;
495 srcp += srcskip;
514 Uint32 *srcp = (Uint32 *)info->s_pixels;
537 src2 = _mm_cvtsi32_si64(*srcp); /* src(ARGB) -> src2 (0000ARGB)*/
552 ++srcp;
560 src1 = *(__m64*)srcp; /* 2 x src -> src1(ARGBARGB)*/
585 srcp += 2;
588 srcp += srcskip;
600 Uint32 *srcp = (Uint32 *)info->s_pixels;
618 Uint32 alpha = *srcp & amask;
623 *dstp = (*srcp & chanmask) | (*dstp & ~chanmask);
625 src1 = _mm_cvtsi32_si64(*srcp); /* src(ARGB) -> src1 (0000ARGB)*/
646 ++srcp;
649 srcp += srcskip;
942 Uint32 *srcp = (Uint32 *)info->s_pixels;
994 RETRIEVE_RGB_PIXEL(((Uint8 *)srcp), 4, Pixel); \
1002 srcp++; \
1008 vector unsigned char valigner = VEC_ALIGNER(srcp);
1009 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1017 /* s = *srcp */
1018 voverflow = (vector unsigned char)vec_ld(15, srcp);
1046 srcp += 4;
1055 srcp += srcskip;
1065 Uint32 *srcp = (Uint32 *)info->s_pixels;
1098 DISEMBLE_RGBA((Uint8 *)srcp, 4, srcfmt, Pixel, sR, sG, sB, sA); \
1104 ++srcp; \
1113 vector unsigned char valigner = VEC_ALIGNER(srcp);
1114 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1121 /* s = *srcp */
1122 voverflow = (vector unsigned char)vec_ld(15, srcp);
1143 srcp += 4;
1151 srcp += srcskip;
1162 Uint32 *srcp = (Uint32 *)info->s_pixels;
1190 Uint32 s = *srcp; \
1207 ++srcp; \
1214 vector unsigned char valigner = VEC_ALIGNER(srcp);
1215 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1222 /* s = *srcp */
1223 voverflow = (vector unsigned char)vec_ld(15, srcp);
1241 srcp += 4;
1248 srcp += srcskip;
1259 Uint32 *srcp = (Uint32 *)info->s_pixels;
1298 DISEMBLE_RGB(((Uint8 *)srcp), 4, srcfmt, Pixel, sR, sG, sB); \
1302 ++srcp; \
1309 vector unsigned char valigner = VEC_ALIGNER(srcp);
1310 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1316 /* s = *srcp */
1317 voverflow = (vector unsigned char)vec_ld(15, srcp);
1334 srcp += 4;
1343 srcp += srcskip;
1355 Uint32 *srcp = (Uint32 *)info->s_pixels;
1379 Uint32 s = *srcp; \
1389 ++srcp; \
1396 vector unsigned char valigner = VEC_ALIGNER(srcp);
1397 vector unsigned char vs = (vector unsigned char)vec_ld(0, srcp);
1403 /* s = *srcp */
1404 voverflow = (vector unsigned char)vec_ld(15, srcp);
1418 srcp += 4;
1427 srcp += srcskip;
1441 Uint32 *srcp = (Uint32 *)info->s_pixels;
1448 Uint32 s = *srcp++;
1453 srcp += srcskip;
1467 Uint32 *srcp = (Uint32 *)info->s_pixels;
1479 s = *srcp;
1489 ++srcp;
1493 s = *srcp;
1501 ((srcp[1] & 0xff00) << 8);
1508 ++srcp;
1510 s1 = *srcp;
1518 ++srcp;
1521 srcp += srcskip;
1532 Uint32 *srcp = (Uint32 *)info->s_pixels;
1543 Uint32 s = *srcp;
1568 ++srcp;
1571 srcp += srcskip;
1582 Uint32 *srcp = (Uint32 *)info->s_pixels;
1619 : : "r" (srcp), "r" (dstp) );
1621 alpha = *srcp & amask;
1640 : : "r" (srcp), "r" (dstp) );
1678 : : "r" (srcp), "r" (dstp), "r" (alpha) );
1681 ++srcp;
1684 srcp += srcskip;
1700 Uint32 *srcp = (Uint32 *)info->s_pixels;
1720 _m_prefetch(srcp + 16);
1723 alpha = *srcp & amask;
1728 *dstp = (*srcp & chanmask) | (*dstp & ~chanmask);
1730 src1 = _mm_cvtsi32_si64(*srcp); /* src(ARGB) -> src1 (0000ARGB)*/
1751 ++srcp;
1754 srcp += srcskip;
1778 Uint16 *srcp = (Uint16 *)info->s_pixels;
1784 if(((uintptr_t)srcp ^ (uintptr_t)dstp) & 2) {
1795 Uint16 d = *dstp, s = *srcp;
1798 srcp++;
1801 srcp++; /* srcp is now 32-bit aligned */
1804 prev_sw = ((Uint32 *)srcp)[-1];
1808 sw = *(Uint32 *)srcp;
1818 srcp += 2;
1831 srcp++;
1834 srcp += srcskip - 1;
1841 if((uintptr_t)srcp & 2) {
1842 Uint16 d = *dstp, s = *srcp;
1844 srcp++;
1848 /* srcp and dstp are now 32-bit aligned */
1851 Uint32 sw = *(Uint32 *)srcp;
1854 srcp += 2;
1861 Uint16 d = *dstp, s = *srcp;
1863 srcp++;
1866 srcp += srcskip;
1882 Uint16 *srcp = (Uint16 *)info->s_pixels;
1908 s = *srcp++;
1921 s = *srcp++;
1933 s = *srcp++;
1946 movq_m2r((*srcp), mm2);/* 4 src pixels -> mm2 */
2002 srcp += 4;
2005 srcp += srcskip;
2021 Uint16 *srcp = (Uint16 *)info->s_pixels;
2047 s = *srcp++;
2060 s = *srcp++;
2072 s = *srcp++;
2085 movq_m2r((*srcp), mm2);/* 4 src pixels -> mm2 */
2146 srcp += 4;
2149 srcp += srcskip;
2167 Uint16 *srcp = (Uint16 *)info->s_pixels;
2192 s = *srcp++;
2205 s = *srcp++;
2217 s = *srcp++;
2230 src1 = *(__m64*)srcp; /* 4 src pixels -> src1 */
2282 srcp += 4;
2285 srcp += srcskip;
2301 Uint16 *srcp = (Uint16 *)info->s_pixels;
2327 s = *srcp++;
2340 s = *srcp++;
2352 s = *srcp++;
2365 src1 = *(__m64*)srcp; /* 4 src pixels -> src1 */
2417 srcp += 4;
2420 srcp += srcskip;
2437 Uint16 *srcp = (Uint16 *)info->s_pixels;
2445 Uint32 s = *srcp++;
2458 srcp += srcskip;
2473 Uint16 *srcp = (Uint16 *)info->s_pixels;
2481 Uint32 s = *srcp++;
2494 srcp += srcskip;
2505 Uint32 *srcp = (Uint32 *)info->s_pixels;
2512 Uint32 s = *srcp;
2535 srcp++;
2538 srcp += srcskip;
2548 Uint32 *srcp = (Uint32 *)info->s_pixels;
2556 Uint32 s = *srcp;
2579 srcp++;
2582 srcp += srcskip;