/external/opencore/codecs_v2/video/avc_h264/enc/src/ |
H A D | sad_inline.h | 37 __inline int32 sad_4pixel(int32 src1, int32 src2, int32 mask) argument 41 x7 = src2 ^ src1; /* check odd/even combination */ 42 if ((uint32)src2 >= (uint32)src1) 44 src1 = src2 - src1; /* subs */ 48 src1 = src1 - src2; 50 x7 = x7 ^ src1; /* only odd bytes need to add carry */ 53 src1 = src1 184 sad_4pixel(int32 src1, int32 src2, int32 mask) argument 190 EOR x7, src2, src1; /* check odd/even combination */ local 191 SUBS src1, src2, src1; local 192 EOR x7, x7, src1; local 196 ADD src1, src1, x7, asr #7; /* add 0xFF to the negative byte, add back carry */ local 197 EOR src1, src1, x7, asr #7; /* take absolute value of negative byte */ local 203 sad_4pixelN(int32 src1, int32 src2, int32 mask) argument 209 EOR x7, src2, src1; /* check odd/even combination */ local 210 ADDS src1, src2, src1; local 211 EOR x7, x7, src1; /* only odd bytes need to add carry */ local 214 SUB src1, src1, x7, asr #7; /* add 0xFF to the negative byte, add back carry */ local 215 EOR src1, src1, x7, asr #7; /* take absolute value of negative byte */ local 347 sad_4pixel(int32 src1, int32 src2, int32 mask) argument 356 sad_4pixelN(int32 src1, int32 src2, int32 mask) argument [all...] |
/external/kernel-headers/original/linux/ |
H A D | bitmap.h | 33 * bitmap_and(dst, src1, src2, nbits) *dst = *src1 & *src2 34 * bitmap_or(dst, src1, src2, nbits) *dst = *src1 | *src2 35 * bitmap_xor(dst, src1, src2, nbits) *dst = *src1 ^ *src2 36 * bitmap_andnot(dst, src1, src2, nbits) *dst = *src1 & ~(*src2) 38 * bitmap_equal(src1, src2, nbits) Are *src1 an 160 bitmap_and(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 169 bitmap_or(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 178 bitmap_xor(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 187 bitmap_andnot(unsigned long *dst, const unsigned long *src1, const unsigned long *src2, int nbits) argument 205 bitmap_equal(const unsigned long *src1, const unsigned long *src2, int nbits) argument 214 bitmap_intersects(const unsigned long *src1, const unsigned long *src2, int nbits) argument 223 bitmap_subset(const unsigned long *src1, const unsigned long *src2, int nbits) argument [all...] |
H A D | cpumask.h | 27 * void cpus_and(dst, src1, src2) dst = src1 & src2 [intersection] 28 * void cpus_or(dst, src1, src2) dst = src1 | src2 [union] 29 * void cpus_xor(dst, src1, src2) dst = src1 ^ src2 30 * void cpus_andnot(dst, src1, src2) dst = src1 & ~src2 124 #define cpus_and(dst, src1, src2) __cpus_and(&(dst), &(src1), [all...] |
H A D | nodemask.h | 27 * void nodes_and(dst, src1, src2) dst = src1 & src2 [intersection] 28 * void nodes_or(dst, src1, src2) dst = src1 | src2 [union] 29 * void nodes_xor(dst, src1, src2) dst = src1 ^ src2 30 * void nodes_andnot(dst, src1, src2) dst = src1 & ~src2 125 #define nodes_and(dst, src1, src2) \ 126 __nodes_and(&(dst), &(src1), [all...] |
/external/opencore/codecs_v2/video/m4v_h263/enc/src/ |
H A D | sad_inline.h | 42 __inline int32 sad_4pixel(int32 src1, int32 src2, int32 mask) argument 46 x7 = src2 ^ src1; /* check odd/even combination */ 47 if ((uint32)src2 >= (uint32)src1) 49 src1 = src2 - src1; /* subs */ 53 src1 = src1 - src2; 55 x7 = x7 ^ src1; /* only odd bytes need to add carry */ 58 src1 = src1 189 sad_4pixel(int32 src1, int32 src2, int32 mask) argument 195 EOR x7, src2, src1; /* check odd/even combination */ local 196 SUBS src1, src2, src1; local 197 EOR x7, x7, src1; local 201 ADD src1, src1, x7, asr #7; /* add 0xFF to the negative byte, add back carry */ local 202 EOR src1, src1, x7, asr #7; /* take absolute value of negative byte */ local 208 sad_4pixelN(int32 src1, int32 src2, int32 mask) argument 214 EOR x7, src2, src1; /* check odd/even combination */ local 215 ADDS src1, src2, src1; local 216 EOR x7, x7, src1; /* only odd bytes need to add carry */ local 219 SUB src1, src1, x7, asr #7; /* add 0xFF to the negative byte, add back carry */ local 220 EOR src1, src1, x7, asr #7; /* take absolute value of negative byte */ local 365 sad_4pixel(int32 src1, int32 src2, int32 mask) argument 390 sad_4pixelN(int32 src1, int32 src2, int32 mask) argument [all...] |
/external/qemu/target-arm/ |
H A D | neon_helper.c | 185 #define NEON_USAT(dest, src1, src2, type) do { \ 186 uint32_t tmp = (uint32_t)src1 + (uint32_t)src2; \ 193 #define NEON_FN(dest, src1, src2) NEON_USAT(dest, src1, src2, uint8_t) 196 #define NEON_FN(dest, src1, src2) NEON_USAT(dest, src1, src2, uint16_t) 201 #define NEON_SSAT(dest, src1, src2, type) do { \ 202 int32_t tmp = (uint32_t)src1 + (uint32_t)src2; \ 213 #define NEON_FN(dest, src1, src2) NEON_SSAT(dest, src1, src 274 neon_hadd_u32(uint32_t src1, uint32_t src2) argument 291 neon_rhadd_s32(int32_t src1, int32_t src2) argument 301 neon_rhadd_u32(uint32_t src1, uint32_t src2) argument 318 neon_hsub_s32(int32_t src1, int32_t src2) argument 328 neon_hsub_u32(uint32_t src1, uint32_t src2) argument [all...] |
H A D | op_helper.c | 481 uint64_t HELPER(neon_add_saturate_s64)(uint64_t src1, uint64_t src2) argument 485 res = src1 + src2; 486 if (((res ^ src1) & SIGNBIT64) && !((src1 ^ src2) & SIGNBIT64)) { 488 res = ((int64_t)src1 >> 63) ^ ~SIGNBIT64; 493 uint64_t HELPER(neon_add_saturate_u64)(uint64_t src1, uint64_t src2) argument 497 res = src1 + src2; 498 if (res < src1) { 505 uint64_t HELPER(neon_sub_saturate_s64)(uint64_t src1, uint64_t src2) argument 509 res = src1 517 neon_sub_saturate_u64(uint64_t src1, uint64_t src2) argument [all...] |
/external/bison/lib/ |
H A D | vbitset.c | 502 vbitset_and (bitset dst, bitset src1, bitset src2) argument 512 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE_ (src2))); 515 ssize1 = VBITSET_SIZE (src1); 518 src1p = VBITSET_WORDS (src1); 529 vbitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 540 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE_ (src2))); 543 ssize1 = VBITSET_SIZE (src1); 546 src1p = VBITSET_WORDS (src1); 582 vbitset_andn (bitset dst, bitset src1, bitset src2) argument 592 vbitset_resize (dst, max (BITSET_SIZE_ (src1), BITSET_SIZE 622 vbitset_andn_cmp(bitset dst, bitset src1, bitset src2) argument 687 vbitset_or(bitset dst, bitset src1, bitset src2) argument 723 vbitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 778 vbitset_xor(bitset dst, bitset src1, bitset src2) argument 814 vbitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 872 vbitset_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 902 vbitset_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 939 vbitset_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 969 vbitset_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 1006 vbitset_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument 1036 vbitset_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | bitset_stats.c | 460 bitset_stats_and (bitset dst, bitset src1, bitset src2) argument 462 BITSET_CHECK3_ (dst, src1, src2); 463 BITSET_AND_ (dst->s.bset, src1->s.bset, src2->s.bset); 468 bitset_stats_and_cmp (bitset dst, bitset src1, bitset src2) argument 470 BITSET_CHECK3_ (dst, src1, src2); 471 return BITSET_AND_CMP_ (dst->s.bset, src1->s.bset, src2->s.bset); 476 bitset_stats_andn (bitset dst, bitset src1, bitset src2) argument 478 BITSET_CHECK3_ (dst, src1, src2); 479 BITSET_ANDN_ (dst->s.bset, src1->s.bset, src2->s.bset); 484 bitset_stats_andn_cmp (bitset dst, bitset src1, bitse argument 492 bitset_stats_or(bitset dst, bitset src1, bitset src2) argument 500 bitset_stats_or_cmp(bitset dst, bitset src1, bitset src2) argument 508 bitset_stats_xor(bitset dst, bitset src1, bitset src2) argument 516 bitset_stats_xor_cmp(bitset dst, bitset src1, bitset src2) argument 524 bitset_stats_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 532 bitset_stats_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 540 bitset_stats_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 548 bitset_stats_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 556 bitset_stats_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument 564 bitset_stats_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | bitset.c | 409 bitset_op4_cmp (bitset dst, bitset src1, bitset src2, bitset src3, argument 428 bitset_or (tmp, src1, src2); 433 bitset_and (tmp, src1, src2); 438 bitset_andn (tmp, src1, src2); 450 bitset_and_or_ (bitset dst, bitset src1, bitset src2, bitset src3) argument 452 bitset_and_or_cmp_ (dst, src1, src2, src3); 459 bitset_and_or_cmp_ (bitset dst, bitset src1, bitset src2, bitset src3) argument 461 return bitset_op4_cmp (dst, src1, src2, src3, BITSET_OP_AND_OR); 467 bitset_andn_or_ (bitset dst, bitset src1, bitset src2, bitset src3) argument 469 bitset_andn_or_cmp_ (dst, src1, src 476 bitset_andn_or_cmp_(bitset dst, bitset src1, bitset src2, bitset src3) argument 484 bitset_or_and_(bitset dst, bitset src1, bitset src2, bitset src3) argument 493 bitset_or_and_cmp_(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | abitset.c | 427 abitset_and (bitset dst, bitset src1, bitset src2) argument 430 bitset_word *src1p = ABITSET_WORDS (src1); 441 abitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 445 bitset_word *src1p = ABITSET_WORDS (src1); 465 abitset_andn (bitset dst, bitset src1, bitset src2) argument 468 bitset_word *src1p = ABITSET_WORDS (src1); 479 abitset_andn_cmp (bitset dst, bitset src1, bitset src2) argument 483 bitset_word *src1p = ABITSET_WORDS (src1); 503 abitset_or (bitset dst, bitset src1, bitset src2) argument 506 bitset_word *src1p = ABITSET_WORDS (src1); 517 abitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 541 abitset_xor(bitset dst, bitset src1, bitset src2) argument 555 abitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 579 abitset_and_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 594 abitset_and_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 619 abitset_andn_or(bitset dst, bitset src1, bitset src2, bitset src3) argument 634 abitset_andn_or_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument 659 abitset_or_and(bitset dst, bitset src1, bitset src2, bitset src3) argument 674 abitset_or_and_cmp(bitset dst, bitset src1, bitset src2, bitset src3) argument [all...] |
H A D | ebitset.c | 1029 ebitset_op3_cmp (bitset dst, bitset src1, bitset src2, enum bitset_ops op) argument 1045 ebitset_resize (dst, max (BITSET_NBITS_ (src1), BITSET_NBITS_ (src2))); 1047 ssize1 = EBITSET_SIZE (src1); 1054 selts1 = EBITSET_ELTS (src1); 1177 ebitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 1188 else if (EBITSET_ZERO_P (src1)) 1195 return ebitset_op3_cmp (dst, src1, src2, BITSET_OP_AND); 1200 ebitset_and (bitset dst, bitset src1, bitset src2) argument 1202 ebitset_and_cmp (dst, src1, src2); 1207 ebitset_andn_cmp (bitset dst, bitset src1, bitse argument 1227 ebitset_andn(bitset dst, bitset src1, bitset src2) argument 1234 ebitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 1249 ebitset_or(bitset dst, bitset src1, bitset src2) argument 1256 ebitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 1271 ebitset_xor(bitset dst, bitset src1, bitset src2) argument [all...] |
H A D | lbitset.c | 1041 lbitset_op3_cmp (bitset dst, bitset src1, bitset src2, enum bitset_ops op) argument 1043 lbitset_elt *selt1 = LBITSET_HEAD (src1); 1199 lbitset_and_cmp (bitset dst, bitset src1, bitset src2) argument 1201 lbitset_elt *selt1 = LBITSET_HEAD (src1); 1219 return lbitset_op3_cmp (dst, src1, src2, BITSET_OP_AND); 1224 lbitset_and (bitset dst, bitset src1, bitset src2) argument 1226 lbitset_and_cmp (dst, src1, src2); 1231 lbitset_andn_cmp (bitset dst, bitset src1, bitset src2) argument 1233 lbitset_elt *selt1 = LBITSET_HEAD (src1); 1239 return lbitset_copy_cmp (dst, src1); 1253 lbitset_andn(bitset dst, bitset src1, bitset src2) argument 1260 lbitset_or_cmp(bitset dst, bitset src1, bitset src2) argument 1278 lbitset_or(bitset dst, bitset src1, bitset src2) argument 1285 lbitset_xor_cmp(bitset dst, bitset src1, bitset src2) argument 1303 lbitset_xor(bitset dst, bitset src1, bitset src2) argument [all...] |
/external/kernel-headers/original/asm-x86/ |
H A D | mpspec_32.h | 50 #define physids_and(dst, src1, src2) bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS) 51 #define physids_or(dst, src1, src2) bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS)
|
/external/v8/src/arm/ |
H A D | assembler-arm.h | 631 void ubfx(Register dst, Register src1, const Operand& src2, 634 void and_(Register dst, Register src1, const Operand& src2, 637 void eor(Register dst, Register src1, const Operand& src2, 640 void sub(Register dst, Register src1, const Operand& src2, 642 void sub(Register dst, Register src1, Register src2, argument 644 sub(dst, src1, Operand(src2), s, cond); 647 void rsb(Register dst, Register src1, const Operand& src2, 650 void add(Register dst, Register src1, const Operand& src2, 653 void adc(Register dst, Register src1, const Operand& src2, 656 void sbc(Register dst, Register src1, cons 663 tst(Register src1, Register src2, Condition cond = al) argument 670 cmp(Register src1, Register src2, Condition cond = al) argument 678 orr(Register dst, Register src1, Register src2, SBit s = LeaveCC, Condition cond = al) argument [all...] |
H A D | assembler-thumb2.h | 631 void ubfx(Register dst, Register src1, const Operand& src2, 634 void and_(Register dst, Register src1, const Operand& src2, 637 void eor(Register dst, Register src1, const Operand& src2, 640 void sub(Register dst, Register src1, const Operand& src2, 642 void sub(Register dst, Register src1, Register src2, argument 644 sub(dst, src1, Operand(src2), s, cond); 647 void rsb(Register dst, Register src1, const Operand& src2, 650 void add(Register dst, Register src1, const Operand& src2, 653 void adc(Register dst, Register src1, const Operand& src2, 656 void sbc(Register dst, Register src1, cons 663 tst(Register src1, Register src2, Condition cond = al) argument 670 cmp(Register src1, Register src2, Condition cond = al) argument 678 orr(Register dst, Register src1, Register src2, SBit s = LeaveCC, Condition cond = al) argument [all...] |
H A D | assembler-arm.cc | 888 void Assembler::ubfx(Register dst, Register src1, const Operand& src2, argument 894 dst.code()*B12 | src2.imm32_*B7 | 0x5*B4 | src1.code()); 898 void Assembler::and_(Register dst, Register src1, const Operand& src2, argument 900 addrmod1(cond | 0*B21 | s, src1, dst, src2); 904 void Assembler::eor(Register dst, Register src1, const Operand& src2, argument 906 addrmod1(cond | 1*B21 | s, src1, dst, src2); 910 void Assembler::sub(Register dst, Register src1, const Operand& src2, argument 912 addrmod1(cond | 2*B21 | s, src1, dst, src2); 916 void Assembler::rsb(Register dst, Register src1, const Operand& src2, argument 918 addrmod1(cond | 3*B21 | s, src1, ds 922 add(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 945 adc(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 951 sbc(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 957 rsc(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 963 tst(Register src1, const Operand& src2, Condition cond) argument 968 teq(Register src1, const Operand& src2, Condition cond) argument 973 cmp(Register src1, const Operand& src2, Condition cond) argument 978 cmn(Register src1, const Operand& src2, Condition cond) argument 983 orr(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 997 bic(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 1009 mla(Register dst, Register src1, Register src2, Register srcA, SBit s, Condition cond) argument 1017 mul(Register dst, Register src1, Register src2, SBit s, Condition cond) argument 1025 smlal(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1038 smull(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1051 umlal(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1064 umull(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1452 vmov(const DwVfpRegister dst, const Register src1, const Register src2, const Condition cond) argument 1538 vadd(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1553 vsub(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1568 vmul(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1583 vdiv(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1598 vcmp(const DwVfpRegister src1, const DwVfpRegister src2, const SBit s, const Condition cond) argument [all...] |
H A D | assembler-thumb2.cc | 868 void Assembler::ubfx(Register dst, Register src1, const Operand& src2, argument 874 dst.code()*B12 | src2.imm32_*B7 | 0x5*B4 | src1.code()); 878 void Assembler::and_(Register dst, Register src1, const Operand& src2, argument 880 addrmod1(cond | 0*B21 | s, src1, dst, src2); 884 void Assembler::eor(Register dst, Register src1, const Operand& src2, argument 886 addrmod1(cond | 1*B21 | s, src1, dst, src2); 890 void Assembler::sub(Register dst, Register src1, const Operand& src2, argument 892 addrmod1(cond | 2*B21 | s, src1, dst, src2); 896 void Assembler::rsb(Register dst, Register src1, const Operand& src2, argument 898 addrmod1(cond | 3*B21 | s, src1, ds 902 add(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 925 adc(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 931 sbc(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 937 rsc(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 943 tst(Register src1, const Operand& src2, Condition cond) argument 948 teq(Register src1, const Operand& src2, Condition cond) argument 953 cmp(Register src1, const Operand& src2, Condition cond) argument 958 cmn(Register src1, const Operand& src2, Condition cond) argument 963 orr(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 977 bic(Register dst, Register src1, const Operand& src2, SBit s, Condition cond) argument 989 mla(Register dst, Register src1, Register src2, Register srcA, SBit s, Condition cond) argument 997 mul(Register dst, Register src1, Register src2, SBit s, Condition cond) argument 1005 smlal(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1018 smull(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1031 umlal(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1044 umull(Register dstL, Register dstH, Register src1, Register src2, SBit s, Condition cond) argument 1430 vmov(const DwVfpRegister dst, const Register src1, const Register src2, const Condition cond) argument 1516 vadd(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1531 vsub(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1546 vmul(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1561 vdiv(const DwVfpRegister dst, const DwVfpRegister src1, const DwVfpRegister src2, const Condition cond) argument 1576 vcmp(const DwVfpRegister src1, const DwVfpRegister src2, const SBit s, const Condition cond) argument [all...] |
/external/v8/src/x64/ |
H A D | macro-assembler-x64.cc | 689 Register src1, 693 if (dst.is(src1)) { 697 // Restore src1. 698 subq(src1, src2); 702 movq(dst, src1); 710 Register src1, 717 if (dst.is(src1)) { 720 movq(dst, src1); 724 } else if (dst.is(src1)) { 728 // Restore src1 914 SmiDiv(Register dst, Register src1, Register src2, Label* on_not_smi_result) argument 977 SmiMod(Register dst, Register src1, Register src2, Label* on_not_smi_result) argument 1046 SmiAnd(Register dst, Register src1, Register src2) argument 1069 SmiOr(Register dst, Register src1, Register src2) argument 1089 SmiXor(Register dst, Register src1, Register src2) argument 1156 SmiShiftLeft(Register dst, Register src1, Register src2, Label* on_not_smi_result) argument 1173 SmiShiftLogicalRight(Register dst, Register src1, Register src2, Label* on_not_smi_result) argument 1209 SmiShiftArithmeticRight(Register dst, Register src1, Register src2) argument 1236 SelectNonSmi(Register dst, Register src1, Register src2, Label* on_not_smis) argument 1349 JumpIfNotBothSmi(Register src1, Register src2, Label* on_not_both_smi) argument 1356 JumpIfNotBothPositiveSmi(Register src1, Register src2, Label* on_not_both_smi) argument [all...] |
H A D | macro-assembler-x64.h | 258 void JumpIfNotBothSmi(Register src1, Register src2, Label* on_not_both_smi); 261 void JumpIfNotBothPositiveSmi(Register src1, Register src2, 306 // If dst is src1, then src1 will be destroyed, even if 309 Register src1, 314 // If dst is src1, then src1 will be destroyed, even if 317 Register src1, 323 // If dst is src1, then src1 wil [all...] |
/external/qemu/distrib/sdl-1.2.12/src/audio/ |
H A D | SDL_mixer.c | 194 Sint16 src1, src2; 201 src1 = ((src[1])<<8|src[0]); 202 ADJUST_VOLUME(src1, volume); 205 dst_sample = src1+src2; 226 Sint16 src1, src2; 233 src1 = ((src[0])<<8|src[1]); 234 ADJUST_VOLUME(src1, volume); 237 dst_sample = src1+src2;
|
/external/webkit/JavaScriptCore/jit/ |
H A D | JITStubs.cpp | 1180 JSValue src1 = stackFrame.args[0].jsValue(); local 1184 bool result = jsLessEq(callFrame, src1, src2); 1608 JSValue src1 = stackFrame.args[0].jsValue(); local 1613 if (src1.getNumber(left) && src2.getNumber(right)) 1617 JSValue result = jsNumber(stackFrame.globalData, src1.toNumber(callFrame) * src2.toNumber(callFrame)); 2033 JSValue src1 = stackFrame.args[0].jsValue(); local 2038 if (src1.getNumber(left) && src2.getNumber(right)) 2042 JSValue result = jsNumber(stackFrame.globalData, src1.toNumber(callFrame) - src2.toNumber(callFrame)); 2329 JSValue src1 = stackFrame.args[0].jsValue(); local 2334 if (src1 2359 JSValue src1 = stackFrame.args[0].jsValue(); local 2372 JSValue src1 = stackFrame.args[0].jsValue(); local 2398 JSValue src1 = stackFrame.args[0].jsValue(); local 2426 JSValue src1 = stackFrame.args[0].jsValue(); local 2553 JSValue src1 = stackFrame.args[0].jsValue(); local 2705 JSValue src1 = stackFrame.args[0].jsValue(); local 2726 JSValue src1 = stackFrame.args[0].jsValue(); local 2887 JSValue src1 = stackFrame.args[0].jsValue(); local 2913 JSValue src1 = stackFrame.args[0].jsValue(); local [all...] |
/external/qemu/distrib/sdl-1.2.12/src/video/ |
H A D | SDL_blit_A.c | 333 punpcklbw_r2r(mm5, mm0); /* low - 0A0R0G0B -> mm0(src1) */ 338 psubw_r2r(mm2, mm0);/* src1 - dst1 -> mm0 */ 456 __m64 src1, src2, dst1, dst2, lmask, hmask, dsta; local 476 src1 = *(__m64*)srcp; /* 2 x src -> src1(ARGBARGB) */ 477 src2 = src1; /* 2 x src -> src2(ARGBARGB) */ 484 dst1 = _mm_and_si64(dst1, src1); /* src & dst -> dst1 */ 520 __m64 src1, src2, dst1, dst2, mm_alpha, mm_zero, dsta; local 559 src1 = *(__m64*)srcp; /* 2 x src -> src1(ARGBARG 609 __m64 src1, dst1, mm_alpha, mm_zero, dmask; local 1709 __m64 src1, dst1, mm_alpha, mm_zero, dmask; local 2172 __m64 src1, dst1, src2, dst2, gmask, bmask, mm_res, mm_alpha; local 2306 __m64 src1, dst1, src2, dst2, rmask, gmask, bmask, mm_res, mm_alpha; local [all...] |
/external/jpeg/ |
H A D | jccolor.c | 323 UINT32 src1 = *in++; local 326 *out0++ = PACK(B0(src0), B3(src0), B2(src1), B1(src2)); 327 *out1++ = PACK(B1(src0), B0(src1), B3(src1), B2(src2)); 328 *out2++ = PACK(B2(src0), B1(src1), B0(src2), B3(src2));
|
/external/webkit/JavaScriptCore/interpreter/ |
H A D | Interpreter.cpp | 1228 /* eq dst(r) src1(r) src2(r) 1230 Checks whether register src1 and register src2 are equal, 1235 JSValue src1 = callFrame->r(vPC[2].u.operand).jsValue(); 1237 if (src1.isInt32() && src2.isInt32()) 1238 callFrame->r(dst) = jsBoolean(src1.asInt32() == src2.asInt32()); 1240 JSValue result = jsBoolean(JSValue::equalSlowCase(callFrame, src1, src2)); 1268 /* neq dst(r) src1(r) src2(r) 1270 Checks whether register src1 and register src2 are not 1275 JSValue src1 = callFrame->r(vPC[2].u.operand).jsValue(); 1277 if (src1 [all...] |