/external/opencv/cxcore/src/ |
H A D | cxcmp.cpp | 57 worktype a1 = _toggle_macro_(src1[x]), \ 67 worktype a1 = _toggle_macro_(src1[x*2]), \ 70 a1 = _toggle_macro_(src1[x*2+1]); \ 81 worktype a1 = _toggle_macro_(src1[x*3]), \ 84 a1 = _toggle_macro_(src1[x*3+1]); \ 88 a1 = _toggle_macro_(src1[x*3+2]); \ 99 worktype a1 = _toggle_macro_(src1[x*4]), \ 102 a1 = _toggle_macro_(src1[x*4+1]); \ 106 a1 = _toggle_macro_(src1[x*4+2]); \ 110 a1 = _toggle_macro_(src1[ 256 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 355 CvMat srcstub1, *src1 = (CvMat*)srcarr; local 567 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 705 CvMat srcstub1, *src1 = (CvMat*)srcarr; local 975 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1076 CvMat srcstub1, *src1 = (CvMat*)srcarr; local 1425 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local [all...] |
H A D | cxlogic.cpp | 63 ( const uchar* src1, int step1, const uchar* src2, int step2, \ 64 uchar* dst, int step, CvSize size ), (src1, step1, src2, step2, dst, step, size) )\ 66 for( ; size.height--; src1 += step1, src2 += step2, dst += step ) \ 70 if( (((size_t)src1 | (size_t)src2 | (size_t)dst) & 3) == 0 ) \ 74 int t0 = __op__(((const int*)(src1+i))[0], ((const int*)(src2+i))[0]);\ 75 int t1 = __op__(((const int*)(src1+i))[1], ((const int*)(src2+i))[1]);\ 80 t0 = __op__(((const int*)(src1+i))[2], ((const int*)(src2+i))[2]); \ 81 t1 = __op__(((const int*)(src1+i))[3], ((const int*)(src2+i))[3]); \ 89 int t = __op__(*(const int*)(src1+i), *(const int*)(src2+i)); \ 96 int t = __op__(((const uchar*)src1)[ 352 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 527 cvXor( const void* src1, const void* src2, void* dst, const void* mask ) argument 544 cvAnd( const void* src1, const void* src2, void* dst, const void* mask ) argument 562 cvOr( const void* src1, const void* src2, void* dst, const void* mask ) argument 573 IPCVAPI_IMPL( CvStatus, icvNot_8u_C1R, ( const uchar* src1, int step1, uchar* dst, int step, CvSize size ), (src1, step1, dst, step, size) ) argument [all...] |
H A D | cxarithm.cpp | 60 worktype t0 = __op__((src1)[i], (src2)[i]); \ 61 worktype t1 = __op__((src1)[i+1], (src2)[i+1]); \ 66 t0 = __op__((src1)[i+2],(src2)[i+2]); \ 67 t1 = __op__((src1)[i+3],(src2)[i+3]); \ 75 worktype t0 = __op__((src1)[i],(src2)[i]); \ 82 ( const type* src1, int step1, const type* src2, int step2, \ 84 (src1, step1, src2, step2, dst, step, size) ) \ 86 step1/=sizeof(src1[0]); step2/=sizeof(src2[0]); step/=sizeof(dst[0]); \ 90 for( ; size.height--; src1 += step1, src2 += step2, dst += step ) \ 92 worktype t0 = __op__((src1)[ 286 CvMat srcstub1, srcstub2, *src1, *src2; local 761 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1321 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1669 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local 1869 icvAddWeighted_8u_fast_C1R( const uchar* src1, int step1, double alpha, const uchar* src2, int step2, double beta, double gamma, uchar* dst, int step, CvSize size ) argument [all...] |
H A D | cxnorm.cpp | 92 worktype t0 = (src1)[x] - (src2)[x];\ 93 worktype t1 = (src1)[x+1]-(src2)[x+1];\ 101 t0 = (src1)[x+2] - (src2)[x+2]; \ 102 t1 = (src1)[x+3] - (src2)[x+3]; \ 113 worktype t0 = (src1)[x] - (src2)[x];\ 122 worktype t0 = (src1)[x*(cn)] - (src2)[x*(cn)]; \ 278 IPCVAPI_IMPL( CvStatus, name,( const arrtype* src1, int step1, \ 280 (src1, step1, src2, step2, size, _norm)) \ 285 step1 /= sizeof(src1[0]); \ 288 for( ; size.height--; src1 [all...] |
H A D | cxcopy.cpp | 304 CvSparseMat* src1 = (CvSparseMat*)src; local 309 dst1->dims = src1->dims; 310 memcpy( dst1->size, src1->size, src1->dims*sizeof(src1->size[0])); 311 dst1->valoffset = src1->valoffset; 312 dst1->idxoffset = src1->idxoffset; 315 if( src1->heap->active_count >= dst1->hashsize*CV_SPARSE_HASH_RATIO ) 318 dst1->hashsize = src1->hashsize; 325 for( node = cvInitSparseMatIterator( src1, 841 const uchar* src1 = src + (size.height - 1)*srcstep; local [all...] |
/external/qemu/target-i386/ |
H A D | helper_template.h | 58 target_long src1, src2; local 59 src1 = CC_SRC; 61 cf = (DATA_TYPE)CC_DST < (DATA_TYPE)src1; 63 af = (CC_DST ^ src1 ^ src2) & 0x10; 66 of = lshift((src1 ^ src2 ^ -1) & (src1 ^ CC_DST), 12 - DATA_BITS) & CC_O; 73 target_long src1; 74 src1 = CC_SRC; 75 cf = (DATA_TYPE)CC_DST < (DATA_TYPE)src1; 82 target_long src1, src [all...] |
/external/kernel-headers/original/linux/ |
H A D | bitmap.h | 33 * bitmap_and(dst, src1, src2, nbits) *dst = *src1 & *src2 34 * bitmap_or(dst, src1, src2, nbits) *dst = *src1 | *src2 35 * bitmap_xor(dst, src1, src2, nbits) *dst = *src1 ^ *src2 36 * bitmap_andnot(dst, src1, src2, nbits) *dst = *src1 & ~(*src2) 38 * bitmap_equal(src1, src2, nbits) Are *src1 an 160 bitmap_and(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 169 bitmap_or(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 178 bitmap_xor(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 187 bitmap_andnot(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 205 bitmap_equal(const unsigned long *src1, const unsigned long *src2, int nbits) argument 214 bitmap_intersects(const unsigned long *src1, const unsigned long *src2, int nbits) argument 223 bitmap_subset(const unsigned long *src1, const unsigned long *src2, int nbits) argument [all...] |
H A D | cpumask.h | 27 * void cpus_and(dst, src1, src2) dst = src1 & src2 [intersection] 28 * void cpus_or(dst, src1, src2) dst = src1 | src2 [union] 29 * void cpus_xor(dst, src1, src2) dst = src1 ^ src2 30 * void cpus_andnot(dst, src1, src2) dst = src1 & ~src2 124 #define cpus_and(dst, src1, src2) __cpus_and(&(dst), &(src1), [all...] |
H A D | nodemask.h | 27 * void nodes_and(dst, src1, src2) dst = src1 & src2 [intersection] 28 * void nodes_or(dst, src1, src2) dst = src1 | src2 [union] 29 * void nodes_xor(dst, src1, src2) dst = src1 ^ src2 30 * void nodes_andnot(dst, src1, src2) dst = src1 & ~src2 125 #define nodes_and(dst, src1, src2) \ 126 __nodes_and(&(dst), &(src1), [all...] |
/external/bison/lib/ |
H A D | vbitset.c | 502 vbitset_and (bitset dst, bitset src1, bitset src2) argument 512 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE_ (src2))); 515 ssize1 = VBITSET_SIZE (src1); 518 src1p = VBITSET_WORDS (src1); 529 vbitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 540 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE_ (src2))); 543 ssize1 = VBITSET_SIZE (src1); 546 src1p = VBITSET_WORDS (src1); 582 vbitset_andn (bitset dst, bitset src1, bitset src2) argument 592 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE 622 vbitset_andn_cmp(bitset dst, bitset src1, bitset src2) argument 687 vbitset_or(bitset dst, bitset src1, bitset src2) argument 723 vbitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 778 vbitset_xor(bitset dst, bitset src1, bitset src2) argument 814 vbitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 872 vbitset_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 902 vbitset_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 939 vbitset_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 969 vbitset_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 1006 vbitset_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument 1036 vbitset_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | bitset_stats.c | 460 bitset_stats_and (bitset dst, bitset src1, bitset src2) argument 462 BITSET_CHECK3_ (dst, src1, src2); 463 BITSET_AND_ (dst->s.bset, src1->s.bset, src2->s.bset); 468 bitset_stats_and_cmp (bitset dst, bitset src1, bitset src2) argument 470 BITSET_CHECK3_ (dst, src1, src2); 471 return BITSET_AND_CMP_ (dst->s.bset, src1->s.bset, src2->s.bset); 476 bitset_stats_andn (bitset dst, bitset src1, bitset src2) argument 478 BITSET_CHECK3_ (dst, src1, src2); 479 BITSET_ANDN_ (dst->s.bset, src1->s.bset, src2->s.bset); 484 bitset_stats_andn_cmp (bitset dst, bitset src1, bitse argument 492 bitset_stats_or(bitset dst, bitset src1, bitset src2) argument 500 bitset_stats_or_cmp(bitset dst, bitset src1, bitset src2) argument 508 bitset_stats_xor(bitset dst, bitset src1, bitset src2) argument 516 bitset_stats_xor_cmp(bitset dst, bitset src1, bitset src2) argument 524 bitset_stats_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 532 bitset_stats_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 540 bitset_stats_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 548 bitset_stats_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 556 bitset_stats_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument 564 bitset_stats_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | bitset.c | 409 bitset_op4_cmp (bitset dst, bitset src1, bitset src2, bitset src3, argument 428 bitset_or (tmp, src1, src2); 433 bitset_and (tmp, src1, src2); 438 bitset_andn (tmp, src1, src2); 450 bitset_and_or_ (bitset dst, bitset src1, bitset src2, bitset src3) argument 452 bitset_and_or_cmp_ (dst, src1, src2, src3); 459 bitset_and_or_cmp_ (bitset dst, bitset src1, bitset src2, bitset src3) argument 461 return bitset_op4_cmp (dst, src1, src2, src3, BITSET_OP_AND_OR); 467 bitset_andn_or_ (bitset dst, bitset src1, bitset src2, bitset src3) argument 469 bitset_andn_or_cmp_ (dst, src1, src 476 bitset_andn_or_cmp_(bitset dst, bitset src1, bitset src2, bitset src3) argument 484 bitset_or_and_(bitset dst, bitset src1, bitset src2, bitset src3) argument 493 bitset_or_and_cmp_(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | abitset.c | 427 abitset_and (bitset dst, bitset src1, bitset src2) argument 430 bitset_word *src1p = ABITSET_WORDS (src1); 441 abitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 445 bitset_word *src1p = ABITSET_WORDS (src1); 465 abitset_andn (bitset dst, bitset src1, bitset src2) argument 468 bitset_word *src1p = ABITSET_WORDS (src1); 479 abitset_andn_cmp (bitset dst, bitset src1, bitset src2) argument 483 bitset_word *src1p = ABITSET_WORDS (src1); 503 abitset_or (bitset dst, bitset src1, bitset src2) argument 506 bitset_word *src1p = ABITSET_WORDS (src1); 517 abitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 541 abitset_xor(bitset dst, bitset src1, bitset src2) argument 555 abitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 579 abitset_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 594 abitset_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 619 abitset_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 634 abitset_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 659 abitset_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument 674 abitset_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | ebitset.c | 1029 ebitset_op3_cmp (bitset dst, bitset src1, bitset src2, enum bitset_ops op) argument 1045 ebitset_resize (dst, max (BITSET_NBITS_ (src1), BITSET_NBITS_ (src2))); 1047 ssize1 = EBITSET_SIZE (src1); 1054 selts1 = EBITSET_ELTS (src1); 1177 ebitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 1188 else if (EBITSET_ZERO_P (src1)) 1195 return ebitset_op3_cmp (dst, src1, src2, BITSET_OP_AND); 1200 ebitset_and (bitset dst, bitset src1, bitset src2) argument 1202 ebitset_and_cmp (dst, src1, src2); 1207 ebitset_andn_cmp (bitset dst, bitset src1, bitse argument 1227 ebitset_andn(bitset dst, bitset src1, bitset src2) argument 1234 ebitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 1249 ebitset_or(bitset dst, bitset src1, bitset src2) argument 1256 ebitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 1271 ebitset_xor(bitset dst, bitset src1, bitset src2) argument [all...] |
H A D | lbitset.c | 1041 lbitset_op3_cmp (bitset dst, bitset src1, bitset src2, enum bitset_ops op) argument 1043 lbitset_elt *selt1 = LBITSET_HEAD (src1); 1199 lbitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 1201 lbitset_elt *selt1 = LBITSET_HEAD (src1); 1219 return lbitset_op3_cmp (dst, src1, src2, BITSET_OP_AND); 1224 lbitset_and (bitset dst, bitset src1, bitset src2) argument 1226 lbitset_and_cmp (dst, src1, src2); 1231 lbitset_andn_cmp (bitset dst, bitset src1, bitset src2) argument 1233 lbitset_elt *selt1 = LBITSET_HEAD (src1); 1239 return lbitset_copy_cmp (dst, src1); 1253 lbitset_andn(bitset dst, bitset src1, bitset src2) argument 1260 lbitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 1278 lbitset_or(bitset dst, bitset src1, bitset src2) argument 1285 lbitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 1303 lbitset_xor(bitset dst, bitset src1, bitset src2) argument [all...] |
/external/qemu/target-arm/ |
H A D | neon_helper.c | 157 #define NEON_USAT(dest, src1, src2, type) do { \ 158 uint32_t tmp = (uint32_t)src1 + (uint32_t)src2; \ 165 #define NEON_FN(dest, src1, src2) NEON_USAT(dest, src1, src2, uint8_t) 168 #define NEON_FN(dest, src1, src2) NEON_USAT(dest, src1, src2, uint16_t) 183 uint64_t HELPER(neon_qadd_u64)(uint64_t src1, uint64_t src2) argument 187 res = src1 + src2; 188 if (res < src1) { 195 #define NEON_SSAT(dest, src1, src 225 neon_qadd_s64(uint64_t src1, uint64_t src2) argument 263 neon_qsub_u64(uint64_t src1, uint64_t src2) argument 306 neon_qsub_s64(uint64_t src1, uint64_t src2) argument 325 neon_hadd_s32(int32_t src1, int32_t src2) argument 335 neon_hadd_u32(uint32_t src1, uint32_t src2) argument 352 neon_rhadd_s32(int32_t src1, int32_t src2) argument 362 neon_rhadd_u32(uint32_t src1, uint32_t src2) argument 379 neon_hsub_s32(int32_t src1, int32_t src2) argument 389 neon_hsub_u32(uint32_t src1, uint32_t src2) argument [all...] |
/external/opencv/cv/src/ |
H A D | _cvmatrix.h | 63 #define icvAddMatrix_32f( src1, src2, dst, w, h ) \ 64 icvAddVector_32f( (src1), (src2), (dst), (w)*(h)) 66 #define icvSubMatrix_32f( src1, src2, dst, w, h ) \ 67 icvSubVector_32f( (src1), (src2), (dst), (w)*(h)) 91 CV_INLINE double icvDotProduct_32f( const float* src1, const float* src2, int len ) argument 94 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i]; 102 CV_INLINE double icvDotProduct_64f( const double* src1, const double* src2, int len ) argument 105 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i]; 113 CV_INLINE void icvMulVectors_32f( const float* src1, const float* src2, argument 118 dst[i] = src1[ 123 icvMulVectors_64d( const double* src1, const double* src2, double* dst, int len ) argument 134 icvAddVector_32f( const float* src1, const float* src2, float* dst, int len ) argument 144 icvAddVector_64d( const double* src1, const double* src2, double* dst, int len ) argument 155 icvSubVector_32f( const float* src1, const float* src2, float* dst, int len ) argument 165 icvSubVector_64d( const double* src1, const double* src2, double* dst, int len ) argument 283 icvMulMatrix_32f( const float* src1, int w1, int h1, const float* src2, int w2, int h2, float* dst ) argument 308 icvMulMatrix_64d( const double* src1, int w1, int h1, const double* src2, int w2, int h2, double* dst ) argument [all...] |
H A D | cvderiv.cpp | 575 const int *src0 = src[-1], *src1 = src[0], *src2 = src[1]; local 580 int s0 = src0[i] - src1[i]*2 + src2[i] + src1[i+width]; 581 int s1 = src0[i+1] - src1[i+1]*2 + src2[i+1] + src1[i+width+1]; 586 dst[i] = (short)(src0[i] - src1[i]*2 + src2[i] + src1[i+width]); 591 int s0 = src0[i] - src1[i]*2 + src2[i] + 592 src0[i+width] + src1[i+width]*2 + src2[i+width]; 593 int s1 = src0[i+1] - src1[ 609 const int *src0 = src[-2], *src1 = src[-1], *src2 = src[0], *src3 = src[1], *src4 = src[2]; local 643 const int* src1 = src[k] + i, *src2 = src[-k] + i; local 659 const int* src1 = src[k] + i, *src2 = src[-k] + i; local 675 const int* src1 = src[k] + i, *src2 = src[-k] + i; local 717 const float *src0 = src[-1], *src1 = src[0], *src2 = src[1]; local 751 const float *src0 = src[-2], *src1 = src[-1], *src2 = src[0], *src3 = src[1], *src4 = src[2]; local 772 const float* src1 = src[k] + i, *src2 = src[-k] + i; local 788 const float* src1 = src[k] + i, *src2 = src[-k] + i; local [all...] |
H A D | cvaccum.cpp | 83 ( const srctype *src1, int step1, const srctype *src2, int step2, \ 85 (src1, step1, src2, step2, dst, dststep, size) ) \ 87 step1 /= sizeof(src1[0]); \ 91 for( ; size.height--; src1 += step1, src2 += step2, dst += dststep ) \ 96 dsttype t0 = dst[x] + cvtmacro(src1[x])*cvtmacro(src2[x]); \ 97 dsttype t1 = dst[x+1] + cvtmacro(src1[x+1])*cvtmacro(src2[x+1]);\ 100 t0 = dst[x + 2] + cvtmacro(src1[x + 2])*cvtmacro(src2[x + 2]); \ 101 t1 = dst[x + 3] + cvtmacro(src1[x + 3])*cvtmacro(src2[x + 3]); \ 106 dst[x] += cvtmacro(src1[x])*cvtmacro(src2[x]); \ 190 ( const srctype *src1, in [all...] |
/external/v8/src/x64/ |
H A D | macro-assembler-x64.h | 332 Register src1, 420 void JumpIfNotBothSmi(Register src1, 426 void JumpUnlessBothNonNegativeSmi(Register src1, Register src2, 480 // If dst is src1, then src1 will be destroyed, even if 484 Register src1, 489 Register src1, 494 Register src1, 498 // If dst is src1, then src1 wil 1298 SmiAdd(Register dst, Register src1, Register src2, LabelType* on_not_smi_result) argument 1318 SmiAdd(Register dst, Register src1, const Operand& src2, LabelType* on_not_smi_result) argument 1338 SmiSub(Register dst, Register src1, Register src2, LabelType* on_not_smi_result) argument 1357 SmiSub(Register dst, Register src1, const Operand& src2, LabelType* on_not_smi_result) argument 1376 SmiMul(Register dst, Register src1, Register src2, LabelType* on_not_smi_result) argument 1521 SmiDiv(Register dst, Register src1, Register src2, LabelType* on_not_smi_result) argument 1585 SmiMod(Register dst, Register src1, Register src2, LabelType* on_not_smi_result) argument 1659 SmiShiftLogicalRight(Register dst, Register src1, Register src2, LabelType* on_not_smi_result) argument 1698 SelectNonSmi(Register dst, Register src1, Register src2, LabelType* on_not_smis) argument 1785 JumpIfNotBothSmi(Register src1, Register src2, LabelType* on_not_both_smi) argument 1794 JumpUnlessBothNonNegativeSmi(Register src1, Register src2, LabelType* on_not_both_smi) argument 1803 SmiOrIfSmis(Register dst, Register src1, Register src2, LabelType* on_not_smis) argument [all...] |
/external/kernel-headers/original/asm-x86/ |
H A D | mpspec_32.h | 50 #define physids_and(dst, src1, src2) bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS) 51 #define physids_or(dst, src1, src2) bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
|
/external/v8/src/arm/ |
H A D | macro-assembler-arm.h | 127 void And(Register dst, Register src1, const Operand& src2, 197 void Push(Register src1, Register src2, Condition cond = al) { argument 198 ASSERT(!src1.is(src2)); 199 if (src1.code() > src2.code()) { 200 stm(db_w, sp, src1.bit() | src2.bit(), cond); 202 str(src1, MemOperand(sp, 4, NegPreIndex), cond); 208 void Push(Register src1, Register src2, Register src3, Condition cond = al) { argument 209 ASSERT(!src1.is(src2)); 211 ASSERT(!src1.is(src3)); 212 if (src1 226 Push(Register src1, Register src2, Register src3, Register src4, Condition cond = al) argument 256 Pop(Register src1, Register src2, Condition cond = al) argument [all...] |
H A D | assembler-arm.h | 756 void and_(Register dst, Register src1, const Operand& src2, 759 void eor(Register dst, Register src1, const Operand& src2, 762 void sub(Register dst, Register src1, const Operand& src2, 764 void sub(Register dst, Register src1, Register src2, argument 766 sub(dst, src1, Operand(src2), s, cond); 769 void rsb(Register dst, Register src1, const Operand& src2, 772 void add(Register dst, Register src1, const Operand& src2, 774 void add(Register dst, Register src1, Register src2, argument 776 add(dst, src1, Operand(src2), s, cond); 779 void adc(Register dst, Register src1, cons 789 tst(Register src1, Register src2, Condition cond = al) argument 796 cmp(Register src1, Register src2, Condition cond = al) argument 805 orr(Register dst, Register src1, Register src2, SBit s = LeaveCC, Condition cond = al) argument [all...] |
/external/opencv/cvaux/src/ |
H A D | cvdpstereo.cpp | 82 void icvFindStereoCorrespondenceByBirchfieldDP( uchar* src1, uchar* src2, argument 115 uchar* srcdata1 = src1 + widthStep * y; 322 if( ( CV_IMAX3( src1[(y-1)*widthStep+x], src1[y*widthStep+x], 323 src1[(y+1)*widthStep+x] ) - 324 CV_IMIN3( src1[(y-1)*widthStep+x], src1[y*widthStep+x], 325 src1[(y+1)*widthStep+x] ) ) >= ICV_BIRCH_DIFF_LUM ) 506 CvMat *src1, *src2; 511 CV_CALL( src1 [all...] |
/external/qemu/distrib/sdl-1.2.12/src/audio/ |
H A D | SDL_mixer.c | 194 Sint16 src1, src2; 201 src1 = ((src[1])<<8|src[0]); 202 ADJUST_VOLUME(src1, volume); 205 dst_sample = src1+src2; 226 Sint16 src1, src2; 233 src1 = ((src[0])<<8|src[1]); 234 ADJUST_VOLUME(src1, volume); 237 dst_sample = src1+src2;
|