Searched refs:src1 (Results 1 - 25 of 98) sorted by relevance

1234

/external/clang/test/CodeGen/
H A Darm-neon-misc.c24 void t2(uint64_t *src1, uint8_t *src2, uint64x2_t *dst) { argument
26 uint64x2_t q = vld1q_u64(src1);
/external/opencv/cxcore/src/
H A Dcxcmp.cpp57 worktype a1 = _toggle_macro_(src1[x]), \
67 worktype a1 = _toggle_macro_(src1[x*2]), \
70 a1 = _toggle_macro_(src1[x*2+1]); \
81 worktype a1 = _toggle_macro_(src1[x*3]), \
84 a1 = _toggle_macro_(src1[x*3+1]); \
88 a1 = _toggle_macro_(src1[x*3+2]); \
99 worktype a1 = _toggle_macro_(src1[x*4]), \
102 a1 = _toggle_macro_(src1[x*4+1]); \
106 a1 = _toggle_macro_(src1[x*4+2]); \
110 a1 = _toggle_macro_(src1[
256 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
355 CvMat srcstub1, *src1 = (CvMat*)srcarr; local
567 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
705 CvMat srcstub1, *src1 = (CvMat*)srcarr; local
975 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
1076 CvMat srcstub1, *src1 = (CvMat*)srcarr; local
1425 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
[all...]
H A Dcxlogic.cpp63 ( const uchar* src1, int step1, const uchar* src2, int step2, \
64 uchar* dst, int step, CvSize size ), (src1, step1, src2, step2, dst, step, size) )\
66 for( ; size.height--; src1 += step1, src2 += step2, dst += step ) \
70 if( (((size_t)src1 | (size_t)src2 | (size_t)dst) & 3) == 0 ) \
74 int t0 = __op__(((const int*)(src1+i))[0], ((const int*)(src2+i))[0]);\
75 int t1 = __op__(((const int*)(src1+i))[1], ((const int*)(src2+i))[1]);\
80 t0 = __op__(((const int*)(src1+i))[2], ((const int*)(src2+i))[2]); \
81 t1 = __op__(((const int*)(src1+i))[3], ((const int*)(src2+i))[3]); \
89 int t = __op__(*(const int*)(src1+i), *(const int*)(src2+i)); \
96 int t = __op__(((const uchar*)src1)[
352 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
527 cvXor( const void* src1, const void* src2, void* dst, const void* mask ) argument
544 cvAnd( const void* src1, const void* src2, void* dst, const void* mask ) argument
562 cvOr( const void* src1, const void* src2, void* dst, const void* mask ) argument
573 IPCVAPI_IMPL( CvStatus, icvNot_8u_C1R, ( const uchar* src1, int step1, uchar* dst, int step, CvSize size ), (src1, step1, dst, step, size) ) argument
[all...]
H A Dcxarithm.cpp60 worktype t0 = __op__((src1)[i], (src2)[i]); \
61 worktype t1 = __op__((src1)[i+1], (src2)[i+1]); \
66 t0 = __op__((src1)[i+2],(src2)[i+2]); \
67 t1 = __op__((src1)[i+3],(src2)[i+3]); \
75 worktype t0 = __op__((src1)[i],(src2)[i]); \
82 ( const type* src1, int step1, const type* src2, int step2, \
84 (src1, step1, src2, step2, dst, step, size) ) \
86 step1/=sizeof(src1[0]); step2/=sizeof(src2[0]); step/=sizeof(dst[0]); \
90 for( ; size.height--; src1 += step1, src2 += step2, dst += step ) \
92 worktype t0 = __op__((src1)[
286 CvMat srcstub1, srcstub2, *src1, *src2; local
761 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
1321 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
1669 CvMat srcstub1, *src1 = (CvMat*)srcarr1; local
1869 icvAddWeighted_8u_fast_C1R( const uchar* src1, int step1, double alpha, const uchar* src2, int step2, double beta, double gamma, uchar* dst, int step, CvSize size ) argument
[all...]
H A Dcxnorm.cpp92 worktype t0 = (src1)[x] - (src2)[x];\
93 worktype t1 = (src1)[x+1]-(src2)[x+1];\
101 t0 = (src1)[x+2] - (src2)[x+2]; \
102 t1 = (src1)[x+3] - (src2)[x+3]; \
113 worktype t0 = (src1)[x] - (src2)[x];\
122 worktype t0 = (src1)[x*(cn)] - (src2)[x*(cn)]; \
278 IPCVAPI_IMPL( CvStatus, name,( const arrtype* src1, int step1, \
280 (src1, step1, src2, step2, size, _norm)) \
285 step1 /= sizeof(src1[0]); \
288 for( ; size.height--; src1
[all...]
/external/qemu/target-i386/
H A Dhelper_template.h58 target_long src1, src2; local
59 src1 = CC_SRC;
61 cf = (DATA_TYPE)CC_DST < (DATA_TYPE)src1;
63 af = (CC_DST ^ src1 ^ src2) & 0x10;
66 of = lshift((src1 ^ src2 ^ -1) & (src1 ^ CC_DST), 12 - DATA_BITS) & CC_O;
73 target_long src1;
74 src1 = CC_SRC;
75 cf = (DATA_TYPE)CC_DST < (DATA_TYPE)src1;
82 target_long src1, src
[all...]
/external/kernel-headers/original/linux/
H A Dbitmap.h33 * bitmap_and(dst, src1, src2, nbits) *dst = *src1 & *src2
34 * bitmap_or(dst, src1, src2, nbits) *dst = *src1 | *src2
35 * bitmap_xor(dst, src1, src2, nbits) *dst = *src1 ^ *src2
36 * bitmap_andnot(dst, src1, src2, nbits) *dst = *src1 & ~(*src2)
38 * bitmap_equal(src1, src2, nbits) Are *src1 an
160 bitmap_and(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument
169 bitmap_or(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument
178 bitmap_xor(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument
187 bitmap_andnot(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument
205 bitmap_equal(const unsigned long *src1, const unsigned long *src2, int nbits) argument
214 bitmap_intersects(const unsigned long *src1, const unsigned long *src2, int nbits) argument
223 bitmap_subset(const unsigned long *src1, const unsigned long *src2, int nbits) argument
[all...]
H A Dcpumask.h27 * void cpus_and(dst, src1, src2) dst = src1 & src2 [intersection]
28 * void cpus_or(dst, src1, src2) dst = src1 | src2 [union]
29 * void cpus_xor(dst, src1, src2) dst = src1 ^ src2
30 * void cpus_andnot(dst, src1, src2) dst = src1 & ~src2
124 #define cpus_and(dst, src1, src2) __cpus_and(&(dst), &(src1),
[all...]
H A Dnodemask.h27 * void nodes_and(dst, src1, src2) dst = src1 & src2 [intersection]
28 * void nodes_or(dst, src1, src2) dst = src1 | src2 [union]
29 * void nodes_xor(dst, src1, src2) dst = src1 ^ src2
30 * void nodes_andnot(dst, src1, src2) dst = src1 & ~src2
125 #define nodes_and(dst, src1, src2) \
126 __nodes_and(&(dst), &(src1),
[all...]
/external/bison/lib/
H A Dvbitset.c502 vbitset_and (bitset dst, bitset src1, bitset src2) argument
512 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE_ (src2)));
515 ssize1 = VBITSET_SIZE (src1);
518 src1p = VBITSET_WORDS (src1);
529 vbitset_and_cmp (bitset dst, bitset src1, bitset src2) argument
540 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE_ (src2)));
543 ssize1 = VBITSET_SIZE (src1);
546 src1p = VBITSET_WORDS (src1);
582 vbitset_andn (bitset dst, bitset src1, bitset src2) argument
592 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE
622 vbitset_andn_cmp(bitset dst, bitset src1, bitset src2) argument
687 vbitset_or(bitset dst, bitset src1, bitset src2) argument
723 vbitset_or_cmp(bitset dst, bitset src1, bitset src2) argument
778 vbitset_xor(bitset dst, bitset src1, bitset src2) argument
814 vbitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument
872 vbitset_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument
902 vbitset_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
939 vbitset_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument
969 vbitset_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
1006 vbitset_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument
1036 vbitset_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
[all...]
H A Dbitset_stats.c460 bitset_stats_and (bitset dst, bitset src1, bitset src2) argument
462 BITSET_CHECK3_ (dst, src1, src2);
463 BITSET_AND_ (dst->s.bset, src1->s.bset, src2->s.bset);
468 bitset_stats_and_cmp (bitset dst, bitset src1, bitset src2) argument
470 BITSET_CHECK3_ (dst, src1, src2);
471 return BITSET_AND_CMP_ (dst->s.bset, src1->s.bset, src2->s.bset);
476 bitset_stats_andn (bitset dst, bitset src1, bitset src2) argument
478 BITSET_CHECK3_ (dst, src1, src2);
479 BITSET_ANDN_ (dst->s.bset, src1->s.bset, src2->s.bset);
484 bitset_stats_andn_cmp (bitset dst, bitset src1, bitse argument
492 bitset_stats_or(bitset dst, bitset src1, bitset src2) argument
500 bitset_stats_or_cmp(bitset dst, bitset src1, bitset src2) argument
508 bitset_stats_xor(bitset dst, bitset src1, bitset src2) argument
516 bitset_stats_xor_cmp(bitset dst, bitset src1, bitset src2) argument
524 bitset_stats_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument
532 bitset_stats_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
540 bitset_stats_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument
548 bitset_stats_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
556 bitset_stats_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument
564 bitset_stats_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
[all...]
H A Dbitset.c409 bitset_op4_cmp (bitset dst, bitset src1, bitset src2, bitset src3, argument
428 bitset_or (tmp, src1, src2);
433 bitset_and (tmp, src1, src2);
438 bitset_andn (tmp, src1, src2);
450 bitset_and_or_ (bitset dst, bitset src1, bitset src2, bitset src3) argument
452 bitset_and_or_cmp_ (dst, src1, src2, src3);
459 bitset_and_or_cmp_ (bitset dst, bitset src1, bitset src2, bitset src3) argument
461 return bitset_op4_cmp (dst, src1, src2, src3, BITSET_OP_AND_OR);
467 bitset_andn_or_ (bitset dst, bitset src1, bitset src2, bitset src3) argument
469 bitset_andn_or_cmp_ (dst, src1, src
476 bitset_andn_or_cmp_(bitset dst, bitset src1, bitset src2, bitset src3) argument
484 bitset_or_and_(bitset dst, bitset src1, bitset src2, bitset src3) argument
493 bitset_or_and_cmp_(bitset dst, bitset src1, bitset src2, bitset src3) argument
[all...]
H A Dabitset.c427 abitset_and (bitset dst, bitset src1, bitset src2) argument
430 bitset_word *src1p = ABITSET_WORDS (src1);
441 abitset_and_cmp (bitset dst, bitset src1, bitset src2) argument
445 bitset_word *src1p = ABITSET_WORDS (src1);
465 abitset_andn (bitset dst, bitset src1, bitset src2) argument
468 bitset_word *src1p = ABITSET_WORDS (src1);
479 abitset_andn_cmp (bitset dst, bitset src1, bitset src2) argument
483 bitset_word *src1p = ABITSET_WORDS (src1);
503 abitset_or (bitset dst, bitset src1, bitset src2) argument
506 bitset_word *src1p = ABITSET_WORDS (src1);
517 abitset_or_cmp(bitset dst, bitset src1, bitset src2) argument
541 abitset_xor(bitset dst, bitset src1, bitset src2) argument
555 abitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument
579 abitset_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument
594 abitset_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
619 abitset_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument
634 abitset_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
659 abitset_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument
674 abitset_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument
[all...]
H A Debitset.c1029 ebitset_op3_cmp (bitset dst, bitset src1, bitset src2, enum bitset_ops op) argument
1045 ebitset_resize (dst, max (BITSET_NBITS_ (src1), BITSET_NBITS_ (src2)));
1047 ssize1 = EBITSET_SIZE (src1);
1054 selts1 = EBITSET_ELTS (src1);
1177 ebitset_and_cmp (bitset dst, bitset src1, bitset src2) argument
1188 else if (EBITSET_ZERO_P (src1))
1195 return ebitset_op3_cmp (dst, src1, src2, BITSET_OP_AND);
1200 ebitset_and (bitset dst, bitset src1, bitset src2) argument
1202 ebitset_and_cmp (dst, src1, src2);
1207 ebitset_andn_cmp (bitset dst, bitset src1, bitse argument
1227 ebitset_andn(bitset dst, bitset src1, bitset src2) argument
1234 ebitset_or_cmp(bitset dst, bitset src1, bitset src2) argument
1249 ebitset_or(bitset dst, bitset src1, bitset src2) argument
1256 ebitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument
1271 ebitset_xor(bitset dst, bitset src1, bitset src2) argument
[all...]
H A Dlbitset.c1041 lbitset_op3_cmp (bitset dst, bitset src1, bitset src2, enum bitset_ops op) argument
1043 lbitset_elt *selt1 = LBITSET_HEAD (src1);
1199 lbitset_and_cmp (bitset dst, bitset src1, bitset src2) argument
1201 lbitset_elt *selt1 = LBITSET_HEAD (src1);
1219 return lbitset_op3_cmp (dst, src1, src2, BITSET_OP_AND);
1224 lbitset_and (bitset dst, bitset src1, bitset src2) argument
1226 lbitset_and_cmp (dst, src1, src2);
1231 lbitset_andn_cmp (bitset dst, bitset src1, bitset src2) argument
1233 lbitset_elt *selt1 = LBITSET_HEAD (src1);
1239 return lbitset_copy_cmp (dst, src1);
1253 lbitset_andn(bitset dst, bitset src1, bitset src2) argument
1260 lbitset_or_cmp(bitset dst, bitset src1, bitset src2) argument
1278 lbitset_or(bitset dst, bitset src1, bitset src2) argument
1285 lbitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument
1303 lbitset_xor(bitset dst, bitset src1, bitset src2) argument
[all...]
/external/qemu/target-arm/
H A Dneon_helper.c157 #define NEON_USAT(dest, src1, src2, type) do { \
158 uint32_t tmp = (uint32_t)src1 + (uint32_t)src2; \
165 #define NEON_FN(dest, src1, src2) NEON_USAT(dest, src1, src2, uint8_t)
168 #define NEON_FN(dest, src1, src2) NEON_USAT(dest, src1, src2, uint16_t)
183 uint64_t HELPER(neon_qadd_u64)(uint64_t src1, uint64_t src2) argument
187 res = src1 + src2;
188 if (res < src1) {
195 #define NEON_SSAT(dest, src1, src
225 neon_qadd_s64(uint64_t src1, uint64_t src2) argument
263 neon_qsub_u64(uint64_t src1, uint64_t src2) argument
306 neon_qsub_s64(uint64_t src1, uint64_t src2) argument
325 neon_hadd_s32(int32_t src1, int32_t src2) argument
335 neon_hadd_u32(uint32_t src1, uint32_t src2) argument
352 neon_rhadd_s32(int32_t src1, int32_t src2) argument
362 neon_rhadd_u32(uint32_t src1, uint32_t src2) argument
379 neon_hsub_s32(int32_t src1, int32_t src2) argument
389 neon_hsub_u32(uint32_t src1, uint32_t src2) argument
[all...]
/external/opencv/cv/src/
H A D_cvmatrix.h63 #define icvAddMatrix_32f( src1, src2, dst, w, h ) \
64 icvAddVector_32f( (src1), (src2), (dst), (w)*(h))
66 #define icvSubMatrix_32f( src1, src2, dst, w, h ) \
67 icvSubVector_32f( (src1), (src2), (dst), (w)*(h))
91 CV_INLINE double icvDotProduct_32f( const float* src1, const float* src2, int len ) argument
94 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i];
102 CV_INLINE double icvDotProduct_64f( const double* src1, const double* src2, int len ) argument
105 for( int i = 0; i < len; i++ ) s += src1[i]*src2[i];
113 CV_INLINE void icvMulVectors_32f( const float* src1, const float* src2, argument
118 dst[i] = src1[
123 icvMulVectors_64d( const double* src1, const double* src2, double* dst, int len ) argument
134 icvAddVector_32f( const float* src1, const float* src2, float* dst, int len ) argument
144 icvAddVector_64d( const double* src1, const double* src2, double* dst, int len ) argument
155 icvSubVector_32f( const float* src1, const float* src2, float* dst, int len ) argument
165 icvSubVector_64d( const double* src1, const double* src2, double* dst, int len ) argument
283 icvMulMatrix_32f( const float* src1, int w1, int h1, const float* src2, int w2, int h2, float* dst ) argument
308 icvMulMatrix_64d( const double* src1, int w1, int h1, const double* src2, int w2, int h2, double* dst ) argument
[all...]
H A Dcvderiv.cpp575 const int *src0 = src[-1], *src1 = src[0], *src2 = src[1]; local
580 int s0 = src0[i] - src1[i]*2 + src2[i] + src1[i+width];
581 int s1 = src0[i+1] - src1[i+1]*2 + src2[i+1] + src1[i+width+1];
586 dst[i] = (short)(src0[i] - src1[i]*2 + src2[i] + src1[i+width]);
591 int s0 = src0[i] - src1[i]*2 + src2[i] +
592 src0[i+width] + src1[i+width]*2 + src2[i+width];
593 int s1 = src0[i+1] - src1[
609 const int *src0 = src[-2], *src1 = src[-1], *src2 = src[0], *src3 = src[1], *src4 = src[2]; local
643 const int* src1 = src[k] + i, *src2 = src[-k] + i; local
659 const int* src1 = src[k] + i, *src2 = src[-k] + i; local
675 const int* src1 = src[k] + i, *src2 = src[-k] + i; local
717 const float *src0 = src[-1], *src1 = src[0], *src2 = src[1]; local
751 const float *src0 = src[-2], *src1 = src[-1], *src2 = src[0], *src3 = src[1], *src4 = src[2]; local
772 const float* src1 = src[k] + i, *src2 = src[-k] + i; local
788 const float* src1 = src[k] + i, *src2 = src[-k] + i; local
[all...]
H A Dcvaccum.cpp83 ( const srctype *src1, int step1, const srctype *src2, int step2, \
85 (src1, step1, src2, step2, dst, dststep, size) ) \
87 step1 /= sizeof(src1[0]); \
91 for( ; size.height--; src1 += step1, src2 += step2, dst += dststep ) \
96 dsttype t0 = dst[x] + cvtmacro(src1[x])*cvtmacro(src2[x]); \
97 dsttype t1 = dst[x+1] + cvtmacro(src1[x+1])*cvtmacro(src2[x+1]);\
100 t0 = dst[x + 2] + cvtmacro(src1[x + 2])*cvtmacro(src2[x + 2]); \
101 t1 = dst[x + 3] + cvtmacro(src1[x + 3])*cvtmacro(src2[x + 3]); \
106 dst[x] += cvtmacro(src1[x])*cvtmacro(src2[x]); \
190 ( const srctype *src1, in
[all...]
/external/v8/src/arm/
H A Dmacro-assembler-arm.h134 void And(Register dst, Register src1, const Operand& src2,
304 void Push(Register src1, Register src2, Condition cond = al) { argument
305 ASSERT(!src1.is(src2));
306 if (src1.code() > src2.code()) {
307 stm(db_w, sp, src1.bit() | src2.bit(), cond);
309 str(src1, MemOperand(sp, 4, NegPreIndex), cond);
315 void Push(Register src1, Register src2, Register src3, Condition cond = al) { argument
316 ASSERT(!src1.is(src2));
318 ASSERT(!src1.is(src3));
319 if (src1
333 Push(Register src1, Register src2, Register src3, Register src4, Condition cond = al) argument
366 Pop(Register src1, Register src2, Condition cond = al) argument
377 Pop(Register src1, Register src2, Register src3, Condition cond = al) argument
395 Pop(Register src1, Register src2, Register src3, Register src4, Condition cond = al) argument
[all...]
H A Dassembler-arm.h786 void and_(Register dst, Register src1, const Operand& src2,
789 void eor(Register dst, Register src1, const Operand& src2,
792 void sub(Register dst, Register src1, const Operand& src2,
794 void sub(Register dst, Register src1, Register src2, argument
796 sub(dst, src1, Operand(src2), s, cond);
799 void rsb(Register dst, Register src1, const Operand& src2,
802 void add(Register dst, Register src1, const Operand& src2,
804 void add(Register dst, Register src1, Register src2, argument
806 add(dst, src1, Operand(src2), s, cond);
809 void adc(Register dst, Register src1, cons
819 tst(Register src1, Register src2, Condition cond = al) argument
826 cmp(Register src1, Register src2, Condition cond = al) argument
835 orr(Register dst, Register src1, Register src2, SBit s = LeaveCC, Condition cond = al) argument
[all...]
/external/kernel-headers/original/asm-x86/
H A Dmpspec_32.h50 #define physids_and(dst, src1, src2) bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
51 #define physids_or(dst, src1, src2) bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
/external/opencv/cvaux/src/
H A Dcvdpstereo.cpp82 void icvFindStereoCorrespondenceByBirchfieldDP( uchar* src1, uchar* src2, argument
115 uchar* srcdata1 = src1 + widthStep * y;
322 if( ( CV_IMAX3( src1[(y-1)*widthStep+x], src1[y*widthStep+x],
323 src1[(y+1)*widthStep+x] ) -
324 CV_IMIN3( src1[(y-1)*widthStep+x], src1[y*widthStep+x],
325 src1[(y+1)*widthStep+x] ) ) >= ICV_BIRCH_DIFF_LUM )
506 CvMat *src1, *src2;
511 CV_CALL( src1
[all...]
/external/v8/src/x64/
H A Dmacro-assembler-x64.h428 Register src1,
519 void JumpIfNotBothSmi(Register src1,
525 void JumpUnlessBothNonNegativeSmi(Register src1, Register src2,
580 // If dst is src1, then src1 will be destroyed, even if
583 Register src1,
588 Register src1,
594 Register src1,
598 // If dst is src1, then src1 wil
[all...]
H A Dmacro-assembler-x64.cc1127 void MacroAssembler::SmiOrIfSmis(Register dst, Register src1, Register src2,
1130 if (dst.is(src1) || dst.is(src2)) {
1131 ASSERT(!src1.is(kScratchRegister));
1133 movq(kScratchRegister, src1);
1138 movq(dst, src1);
1304 void MacroAssembler::JumpIfNotBothSmi(Register src1,
1308 Condition both_smi = CheckBothSmi(src1, src2);
1313 void MacroAssembler::JumpUnlessBothNonNegativeSmi(Register src1,
1317 Condition both_smi = CheckBothNonNegativeSmi(src1, src2);
1516 Register src1,
[all...]

Completed in 8110 milliseconds

1234