/external/pcre/dist2/src/sljit/ |
H A D | sljitNativeMIPS_common.c | 899 sljit_s32 src1, sljit_sw src1w, 939 if (!(flags & SRC2_IMM) && (flags & CUMULATIVE_OP) && (src1 & SLJIT_IMM) && src1w) { 946 src1 = src2; 955 if (FAST_IS_REG(src1)) { 956 src1_r = src1; 959 else if (src1 & SLJIT_IMM) { 968 if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w)) 1005 if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) { 1006 FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, src1, src1 897 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1183 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1363 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1453 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1736 sljit_emit_cmp(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1857 sljit_emit_fcmp(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument [all...] |
H A D | sljitNativePPC_32.c | 45 sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) 52 SLJIT_ASSERT(src1 == TMP_REG1); 59 SLJIT_ASSERT(src1 == TMP_REG1); 74 SLJIT_ASSERT(src1 == TMP_REG1); 86 SLJIT_ASSERT(src1 == TMP_REG1); 90 SLJIT_ASSERT(src1 == TMP_REG1); 94 SLJIT_ASSERT(src1 == TMP_REG1); 101 return push_inst(compiler, ADDI | D(dst) | A(src1) | compiler->imm); 106 return push_inst(compiler, ADDIS | D(dst) | A(src1) | compiler->imm); 110 return push_inst(compiler, ADDIC | D(dst) | A(src1) | compile 44 emit_single_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) argument [all...] |
H A D | sljitNativePPC_64.c | 133 FAIL_IF(push_inst(compiler, EXTSW | S(src1) | A(TMP_REG1))); \ 134 src1 = TMP_REG1; \ 144 FAIL_IF(push_inst(compiler, EXTSW | S(src1) | A(TMP_REG1))); \ 145 src1 = TMP_REG1; \ 149 sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) 154 SLJIT_ASSERT(src1 == TMP_REG1); 161 SLJIT_ASSERT(src1 == TMP_REG1); 174 SLJIT_ASSERT(src1 == TMP_REG1); 189 SLJIT_ASSERT(src1 == TMP_REG1); 201 SLJIT_ASSERT(src1 148 emit_single_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_s32 src1, sljit_s32 src2) argument [all...] |
H A D | sljitNativePPC_common.c | 1123 sljit_s32 src1, sljit_sw src1w, 1166 if (FAST_IS_REG(src1)) { 1167 src1_r = src1; 1170 else if (src1 & SLJIT_IMM) { 1174 else if (getput_arg_fast(compiler, input_flags | LOAD_DATA, TMP_REG1, src1, src1w)) { 1202 if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) { 1203 FAIL_IF(getput_arg(compiler, input_flags | LOAD_DATA, TMP_REG2, src2, src2w, src1, src1w)); 1204 FAIL_IF(getput_arg(compiler, input_flags | LOAD_DATA, TMP_REG1, src1, src1w, dst, dstw)); 1207 FAIL_IF(getput_arg(compiler, input_flags | LOAD_DATA, TMP_REG1, src1, src1 1121 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 input_flags, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1455 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1862 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1931 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument [all...] |
H A D | sljitNativeSPARC_32.c | 39 sljit_s32 dst, sljit_s32 src1, sljit_sw src2) 48 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 55 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 68 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 78 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 82 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 97 return push_inst(compiler, ADD | (flags & SET_FLAGS) | D(dst) | S1(src1) | ARG2(flags, src2), DR(dst) | (flags & SET_FLAGS)); 100 return push_inst(compiler, ADDC | (flags & SET_FLAGS) | D(dst) | S1(src1) | ARG2(flags, src2), DR(dst) | (flags & SET_FLAGS)); 103 return push_inst(compiler, SUB | (flags & SET_FLAGS) | D(dst) | S1(src1) | ARG2(flags, src2), DR(dst) | (flags & SET_FLAGS)); 106 return push_inst(compiler, SUBC | (flags & SET_FLAGS) | D(dst) | S1(src1) | ARG 38 emit_single_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_s32 src1, sljit_sw src2) argument [all...] |
H A D | sljitNativeSPARC_common.c | 638 sljit_s32 src1, sljit_sw src1w, 675 if (!(flags & SRC2_IMM) && (flags & CUMULATIVE_OP) && (src1 & SLJIT_IMM) && src1w) { 681 src1 = src2; 690 if (FAST_IS_REG(src1)) 691 src1_r = src1; 692 else if (src1 & SLJIT_IMM) { 701 if (getput_arg_fast(compiler, flags | LOAD_DATA, TMP_REG1, src1, src1w)) 738 if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) { 739 FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, TMP_REG2, src2, src2w, src1, src1 636 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 879 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1015 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1094 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument [all...] |
H A D | sljitNativeTILEGX_64.c | 1585 static SLJIT_INLINE sljit_s32 emit_single_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_s32 src1, sljit_sw src2) argument 1592 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 1599 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 1614 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 1629 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 1643 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 1652 SLJIT_ASSERT(src1 == TMP_REG1 && !(flags & SRC2_IMM)); 1663 FAIL_IF(SHRUI(TMP_EREG1, reg_map[src1], 63)); 1669 FAIL_IF(ADDLI(EQUAL_FLAG, reg_map[src1], src2)); 1673 FAIL_IF(ORI(ULESS_FLAG ,reg_map[src1], src 1959 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2260 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2500 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument [all...] |
H A D | sljitNativeX86_common.c | 594 sljit_s32 src1, sljit_sw src1w, 600 sljit_s32 src1, sljit_sw src1w, 1483 sljit_s32 src1, sljit_sw src1w, 1489 EMIT_MOV(compiler, TMP_REG1, 0, src1, src1w); 1501 if (dst == src1 && dstw == src1w) { 1536 if (src1 & SLJIT_IMM) { 1549 inst = emit_x86_instruction(compiler, 1, dst, dstw, src1, src1w); 1553 else if (FAST_IS_REG(src1)) { 1554 inst = emit_x86_instruction(compiler, 1, src1, src1w, dst, dstw); 1559 EMIT_MOV(compiler, TMP_REG1, 0, src1, src1 1480 emit_cum_binary(struct sljit_compiler *compiler, sljit_u8 op_rm, sljit_u8 op_mr, sljit_u8 op_imm, sljit_u8 op_eax_imm, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1596 emit_non_cum_binary(struct sljit_compiler *compiler, sljit_u8 op_rm, sljit_u8 op_mr, sljit_u8 op_imm, sljit_u8 op_eax_imm, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1678 emit_mul(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1808 emit_lea_binary(struct sljit_compiler *compiler, sljit_s32 keep_flags, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1867 emit_cmp_binary(struct sljit_compiler *compiler, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 1918 emit_test_binary(struct sljit_compiler *compiler, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2028 emit_shift(struct sljit_compiler *compiler, sljit_u8 mode, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2117 emit_shift_with_flags(struct sljit_compiler *compiler, sljit_u8 mode, sljit_s32 set_flags, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2152 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2405 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument 2483 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 dst, sljit_sw dstw, sljit_s32 src1, sljit_sw src1w, sljit_s32 src2, sljit_sw src2w) argument [all...] |
/external/opencv/cv/include/ |
H A D | cvcompat.h | 363 #define cvmAdd( src1, src2, dst ) cvAdd( src1, src2, dst, 0 ) 364 #define cvmSub( src1, src2, dst ) cvSub( src1, src2, dst, 0 ) 366 #define cvmMul( src1, src2, dst ) cvMatMulAdd( src1, src2, 0, dst )
|
/external/opencv/cv/src/ |
H A D | _cvipp.h | 111 ( const arrtype* src1, int srcstep1, const arrtype* src2, int srcstep2, \ 128 ( const arrtype* src1, int srcstep1, const arrtype* src2, int srcstep2, \ 145 ( const arrtype* src1, int srcstep1, const arrtype* src2, int srcstep2, \ 619 ( const uchar* src1, int srcstep1, uchar scalar, 623 ( const uchar* src1, int srcstep1, uchar scalar,
|
H A D | _cvmatrix.h | 63 #define icvAddMatrix_32f( src1, src2, dst, w, h ) \ 64 icvAddVector_32f( (src1), (src2), (dst), (w)*(h)) 66 #define icvSubMatrix_32f( src1, src2, dst, w, h ) \ 67 icvSubVector_32f( (src1), (src2), (dst), (w)*(h)) 91 CV_INLINE double icvDotProduct_32f( const float* src1, const float* src2, int len ) argument 94 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i]; 102 CV_INLINE double icvDotProduct_64f( const double* src1, const double* src2, int len ) argument 105 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i]; 113 CV_INLINE void icvMulVectors_32f( const float* src1, const float* src2, argument 118 dst[i] = src1[ 123 icvMulVectors_64d( const double* src1, const double* src2, double* dst, int len ) argument 134 icvAddVector_32f( const float* src1, const float* src2, float* dst, int len ) argument 144 icvAddVector_64d( const double* src1, const double* src2, double* dst, int len ) argument 155 icvSubVector_32f( const float* src1, const float* src2, float* dst, int len ) argument 165 icvSubVector_64d( const double* src1, const double* src2, double* dst, int len ) argument 283 icvMulMatrix_32f( const float* src1, int w1, int h1, const float* src2, int w2, int h2, float* dst ) argument 308 icvMulMatrix_64d( const double* src1, int w1, int h1, const double* src2, int w2, int h2, double* dst ) argument [all...] |
H A D | cvaccum.cpp | 83 ( const srctype *src1, int step1, const srctype *src2, int step2, \ 85 (src1, step1, src2, step2, dst, dststep, size) ) \ 87 step1 /= sizeof(src1[0]); \ 91 for( ; size.height--; src1 += step1, src2 += step2, dst += dststep ) \ 96 dsttype t0 = dst[x] + cvtmacro(src1[x])*cvtmacro(src2[x]); \ 97 dsttype t1 = dst[x+1] + cvtmacro(src1[x+1])*cvtmacro(src2[x+1]);\ 100 t0 = dst[x + 2] + cvtmacro(src1[x + 2])*cvtmacro(src2[x + 2]); \ 101 t1 = dst[x + 3] + cvtmacro(src1[x + 3])*cvtmacro(src2[x + 3]); \ 106 dst[x] += cvtmacro(src1[x])*cvtmacro(src2[x]); \ 190 ( const srctype *src1, in [all...] |
H A D | cvcolor.cpp | 434 const arrtype* src1 = src + i*3; \ 437 status = ipp_func( src1, CV_STUB_STEP, \ 983 const uchar* src1 = src + i*3; local 988 float a = CV_8TO32F(src1[k])*pre_coeffs[0] + pre_coeffs[1]; 989 float b = CV_8TO32F(src1[k+1])*pre_coeffs[2] + pre_coeffs[3]; 990 float c = CV_8TO32F(src1[k+2])*pre_coeffs[4] + pre_coeffs[5]; 1376 const uchar* src1 = src + i*3; local 1380 uchar h = icvHue180To255[src1[k]]; 1381 uchar s = src1[k+1]; 1382 uchar v = src1[ [all...] |
H A D | cvderiv.cpp | 575 const int *src0 = src[-1], *src1 = src[0], *src2 = src[1]; local 580 int s0 = src0[i] - src1[i]*2 + src2[i] + src1[i+width]; 581 int s1 = src0[i+1] - src1[i+1]*2 + src2[i+1] + src1[i+width+1]; 586 dst[i] = (short)(src0[i] - src1[i]*2 + src2[i] + src1[i+width]); 591 int s0 = src0[i] - src1[i]*2 + src2[i] + 592 src0[i+width] + src1[i+width]*2 + src2[i+width]; 593 int s1 = src0[i+1] - src1[ 609 const int *src0 = src[-2], *src1 = src[-1], *src2 = src[0], *src3 = src[1], *src4 = src[2]; local 643 const int* src1 = src[k] + i, *src2 = src[-k] + i; local 659 const int* src1 = src[k] + i, *src2 = src[-k] + i; local 675 const int* src1 = src[k] + i, *src2 = src[-k] + i; local 717 const float *src0 = src[-1], *src1 = src[0], *src2 = src[1]; local 751 const float *src0 = src[-2], *src1 = src[-1], *src2 = src[0], *src3 = src[1], *src4 = src[2]; local 772 const float* src1 = src[k] + i, *src2 = src[-k] + i; local 788 const float* src1 = src[k] + i, *src2 = src[-k] + i; local [all...] |
H A D | cvfilter.cpp | 1357 const int *src0 = src[-1], *src1 = src[0], *src2 = src[1]; 1361 int s0 = src0[i] + src1[i]*2 + src2[i], 1362 s1 = src0[i+1] + src1[i+1]*2 + src2[i+1]; 1369 const int *src0 = src[-1], *src1 = src[0], *src2 = src[1]; 1373 int s0 = src1[i]*10 + (src0[i] + src2[i])*3, 1374 s1 = src1[i+1]*10 + (src0[i+1] + src2[i+1])*3; 1682 const float *src0 = src[-1], *src1 = src[0], *src2 = src[1]; 1686 float s0 = src0[i] + src1[i]*2 + src2[i], 1687 s1 = src0[i+1] + src1[i+1]*2 + src2[i+1], 1688 s2 = src0[i+2] + src1[ [all...] |
H A D | cvsmooth.cpp | 826 const uchar* src1 = src0 + src_step; 827 const uchar* src2 = src1 + src_step; 829 src0 = src1; 831 src2 = src1; 840 int p3 = src1[x0], p4 = src1[x1], p5 = src1[x2]; 859 int p3 = src1[x-cn], p4 = src1[x], p5 = src1[ [all...] |
/external/opencv/cvaux/src/ |
H A D | cvdpstereo.cpp | 84 void icvFindStereoCorrespondenceByBirchfieldDP( uchar* src1, uchar* src2, argument 117 uchar* srcdata1 = src1 + widthStep * y; 324 if( ( CV_IMAX3( src1[(y-1)*widthStep+x], src1[y*widthStep+x], 325 src1[(y+1)*widthStep+x] ) - 326 CV_IMIN3( src1[(y-1)*widthStep+x], src1[y*widthStep+x], 327 src1[(y+1)*widthStep+x] ) ) >= ICV_BIRCH_DIFF_LUM ) 508 CvMat *src1, *src2; 513 CV_CALL( src1 [all...] |
/external/opencv/cxcore/include/ |
H A D | cxcore.h | 460 CVAPI(void) cvMerge( const CvArr* src0, const CvArr* src1, 506 /* dst(mask) = src1(mask) + src2(mask) */ 507 CVAPI(void) cvAdd( const CvArr* src1, const CvArr* src2, CvArr* dst, 514 /* dst(mask) = src1(mask) - src2(mask) */ 515 CVAPI(void) cvSub( const CvArr* src1, const CvArr* src2, CvArr* dst, 530 /* dst(idx) = src1(idx) * src2(idx) * scale 532 CVAPI(void) cvMul( const CvArr* src1, const CvArr* src2, 536 dst(idx) = src1(idx) * scale / src2(idx) 537 or dst(idx) = scale / src2(idx) if src1 == 0 */ 538 CVAPI(void) cvDiv( const CvArr* src1, cons [all...] |
/external/opencv/cxcore/src/ |
H A D | _cxipp.h | 93 ( const uchar* src1, int srcstep1, const uchar* src2, int srcstep2, \ 97 ( const ushort* src1, int srcstep1, const ushort* src2, int srcstep2,\ 101 ( const short* src1, int srcstep1, const short* src2, int srcstep2, \ 105 ( const int* src1, int srcstep1, const int* src2, int srcstep2, \ 109 ( const float* src1, int srcstep1, const float* src2, int srcstep2, \ 113 ( const double* src1, int srcstep1, const double* src2, int srcstep2,\ 129 ( const uchar* src1, int srcstep1, const uchar* src2, int srcstep2, \ 431 ( const arrtype* src1, int srcstep1, \ 449 ( const arrtype* src1, int srcstep1, const arrtype* src2, int srcstep2, \ 453 ( const arrtype* src1, in [all...] |
H A D | cxarithm.cpp | 60 worktype t0 = __op__((src1)[i], (src2)[i]); \ 61 worktype t1 = __op__((src1)[i+1], (src2)[i+1]); \ 66 t0 = __op__((src1)[i+2],(src2)[i+2]); \ 67 t1 = __op__((src1)[i+3],(src2)[i+3]); \ 75 worktype t0 = __op__((src1)[i],(src2)[i]); \ 82 ( const type* src1, int step1, const type* src2, int step2, \ 84 (src1, step1, src2, step2, dst, step, size) ) \ 86 step1/=sizeof(src1[0]); step2/=sizeof(src2[0]); step/=sizeof(dst[0]); \ 90 for( ; size.height--; src1 += step1, src2 += step2, dst += step ) \ 92 worktype t0 = __op__((src1)[ 286 CvMat srcstub1, srcstub2, *src1, *src2; local 761 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1321 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1669 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1869 icvAddWeighted_8u_fast_C1R( const uchar* src1, int step1, double alpha, const uchar* src2, int step2, double beta, double gamma, uchar* dst, int step, CvSize size ) argument [all...] |
H A D | cxcmp.cpp | 57 worktype a1 = _toggle_macro_(src1[x]), \ 67 worktype a1 = _toggle_macro_(src1[x*2]), \ 70 a1 = _toggle_macro_(src1[x*2+1]); \ 81 worktype a1 = _toggle_macro_(src1[x*3]), \ 84 a1 = _toggle_macro_(src1[x*3+1]); \ 88 a1 = _toggle_macro_(src1[x*3+2]); \ 99 worktype a1 = _toggle_macro_(src1[x*4]), \ 102 a1 = _toggle_macro_(src1[x*4+1]); \ 106 a1 = _toggle_macro_(src1[x*4+2]); \ 110 a1 = _toggle_macro_(src1[ 256 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 355 CvMat srcstub1, *src1 = (CvMat*)srcarr; local 567 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 705 CvMat srcstub1, *src1 = (CvMat*)srcarr; local 975 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1076 CvMat srcstub1, *src1 = (CvMat*)srcarr; local 1425 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local [all...] |
H A D | cxcopy.cpp | 304 CvSparseMat* src1 = (CvSparseMat*)src; local 309 dst1->dims = src1->dims; 310 memcpy( dst1->size, src1->size, src1->dims*sizeof(src1->size[0])); 311 dst1->valoffset = src1->valoffset; 312 dst1->idxoffset = src1->idxoffset; 315 if( src1->heap->active_count >= dst1->hashsize*CV_SPARSE_HASH_RATIO ) 318 dst1->hashsize = src1->hashsize; 325 for( node = cvInitSparseMatIterator( src1, 841 const uchar* src1 = src + (size.height - 1)*srcstep; local [all...] |
H A D | cxdxt.cpp | 1661 const int* src1 = (const int*)_src1; 1669 t0 = src0[i]; t1 = src1[i]; 1679 t0 = src1[i]; t1 = src1[i+1]; 1691 t0 = src1[i]; t1 = src1[i+1]; 1693 t0 = src1[i+2]; t1 = src1[i+3]; 2460 const datatype* src1 = src + (n-1)*src_step; \ 2471 src += src_step, src1 [all...] |
H A D | cxlogic.cpp | 63 ( const uchar* src1, int step1, const uchar* src2, int step2, \ 64 uchar* dst, int step, CvSize size ), (src1, step1, src2, step2, dst, step, size) )\ 66 for( ; size.height--; src1 += step1, src2 += step2, dst += step ) \ 70 if( (((size_t)src1 | (size_t)src2 | (size_t)dst) & 3) == 0 ) \ 74 int t0 = __op__(((const int*)(src1+i))[0], ((const int*)(src2+i))[0]);\ 75 int t1 = __op__(((const int*)(src1+i))[1], ((const int*)(src2+i))[1]);\ 80 t0 = __op__(((const int*)(src1+i))[2], ((const int*)(src2+i))[2]); \ 81 t1 = __op__(((const int*)(src1+i))[3], ((const int*)(src2+i))[3]); \ 89 int t = __op__(*(const int*)(src1+i), *(const int*)(src2+i)); \ 96 int t = __op__(((const uchar*)src1)[ 352 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 527 cvXor( const void* src1, const void* src2, void* dst, const void* mask ) argument 544 cvAnd( const void* src1, const void* src2, void* dst, const void* mask ) argument 562 cvOr( const void* src1, const void* src2, void* dst, const void* mask ) argument 573 IPCVAPI_IMPL( CvStatus, icvNot_8u_C1R, ( const uchar* src1, int step1, uchar* dst, int step, CvSize size ), (src1, step1, dst, step, size) ) argument [all...] |
H A D | cxmatmul.cpp | 571 typedef CvStatus (CV_STDCALL *CvGEMMSingleMulFunc)( const void* src1, size_t step1, 576 typedef CvStatus (CV_STDCALL *CvGEMMBlockMulFunc)( const void* src1, size_t step1, 580 typedef CvStatus (CV_STDCALL *CvGEMMStoreFunc)( const void* src1, size_t step1, 2118 #define ICV_DEF_MULADDC_CASE_C1( arrtype, temptype, src1, src2, dst, len ) \ 2124 temptype t0 = (src1)[i]*s0 + (src2)[i]; \ 2125 temptype t1 = (src1)[i+1]*s0 + (src2)[i+1]; \ 2130 t0 = (src1)[i+2]*s0 + (src2)[i+2]; \ 2131 t1 = (src1)[i+3]*s0 + (src2)[i+3]; \ 2139 temptype t0 = (src1)[i]*s0 + (src2)[i]; \ 2145 #define ICV_DEF_MULADDC_CASE_C2( arrtype, temptype, src1, src [all...] |