/external/valgrind/VEX/priv/ |
H A D | guest_tilegx_toIR.c | 110 static IRExpr *mkU64 ( ULong i ) function 349 putPC(mkU64(guest_PC_curr_instr + 24)); 363 putPC(mkU64(guest_PC_curr_instr + 8)); 371 putIReg(55, mkU64(guest_PC_curr_instr + 24)); 388 mkU64(guest_PC_curr_instr))); 390 mkU64(24))); 393 putPC(mkU64(guest_PC_curr_instr + 24)); 492 assign(t2, mkU64(extend_s_8to64(imm))); 497 assign(t2, mkU64(extend_s_16to64(imm))); 542 mkU64(extend_s_8to6 [all...] |
H A D | guest_s390_helpers.c | 1810 #define mkU64(v) IRExpr_Const(IRConst_U64(v)) macro 1957 return unop(Iop_1Uto32, binop(Iop_CmpEQ64, cc_dep1, mkU64(0))); 1960 return unop(Iop_1Uto32, binop(Iop_CmpNE64, cc_dep1, mkU64(0))); 1969 return unop(Iop_1Uto32, binop(Iop_CmpLE64S, cc_dep1, mkU64(0))); 1973 return unop(Iop_1Uto32, binop(Iop_CmpLT64S, mkU64(0), cc_dep1)); 1981 mkU64(1))); 2010 return unop(Iop_1Uto32, binop(Iop_CmpEQ64, cc_dep1, mkU64(0))); 2013 return unop(Iop_1Uto32, binop(Iop_CmpNE64, cc_dep1, mkU64(0))); 2116 mkU64(0))); 2121 mkU64( [all...] |
H A D | ir_inject.c | 40 #define mkU64(v) IRExpr_Const(IRConst_U64(v)) macro 90 addr = mkU64(haddr); 91 next_addr = binop(Iop_Add64, addr, mkU64(8)); 151 addr = mkU64(haddr); 152 next_addr = binop(Iop_Add64, addr, mkU64(8));
|
H A D | guest_s390_toIR.c | 207 mkU64(ULong value) function 364 mkU64(guest_IA_curr_instr))); 526 mkIRExprVec_2(mkU64((ULong)text), value)); 587 op = mkU64(opc); 589 dep2 = mkU64(0); 590 ndep = mkU64(0); 601 op = mkU64(opc); 604 ndep = mkU64(0); 626 op = mkU64(opc); 649 stmt(IRStmt_Put(dep1_off, mkU64( [all...] |
H A D | guest_mips_toIR.c | 470 mkU64(extend_s_16to64(imm)))); \ 481 assign(t2, binop(Iop_And64, mkexpr(t1), mkU64(0xFFFFFFFFFFFFFFFCULL))); \ 484 mkexpr(t1), mkU64(0x3)))); 488 assign(t2, binop(Iop_And64, mkexpr(t1), mkU64(0xFFFFFFFFFFFFFFF8ULL))); \ 490 assign(t4, binop(Iop_And64, mkexpr(t1), mkU64(0x7))); 532 putIReg(rt, binop(op, getIReg(rs), mkU64(imm))); 869 static IRExpr *mkU64(ULong i) function 1018 stmt(IRStmt_Put(OFFB_PC, mkU64(d64))); 1058 return mode64 ? mkU64(0x0) : mkU32(0x0); 1097 mkU64( [all...] |
H A D | guest_amd64_helpers.c | 997 # define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 1032 mkU64(0))); 1058 mkU64(1)); 1124 mkU64(1)); 1144 mkU64(1)); 1201 mkU64(1)); 1271 mkU64(1)); 1283 mkU64(1)), 1284 mkU64(1)); 1375 binop(Iop_And64, cc_dep1, mkU64( 1750 # undef mkU64 macro [all...] |
H A D | guest_ppc_toIR.c | 646 static IRExpr* mkU64 ( ULong i ) function 1039 return ty == Ity_I64 ? mkU64(imm64) : mkU32((UInt)imm64); 1056 mkU64(extend_s_16to64(imm16)) : 1065 mkU64(extend_s_32to64(imm32)) : 1564 assign( mask[0], mkU64( 0x5555555555555555ULL ) ); 1565 assign( mask[1], mkU64( 0x3333333333333333ULL ) ); 1566 assign( mask[2], mkU64( 0x0F0F0F0F0F0F0F0FULL ) ); 1567 assign( mask[3], mkU64( 0x00FF00FF00FF00FFULL ) ); 1568 assign( mask[4], mkU64( 0x0000FFFF0000FFFFULL ) ); 1569 assign( mask[5], mkU64( [all...] |
H A D | guest_arm64_toIR.c | 248 static IRExpr* mkU64 ( ULong i ) function 442 //ZZ return binop(Iop_64HLtoV128, mkU64(i), mkU64(i)); 790 case Ity_I64: return mkU64(imm); 1227 return mkU64(0); 1763 return mk_arm64g_calculate_condition_dyn( mkU64(cond << 4) ); 1847 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(cc_op) )); 1869 assign(z64, mkU64(0)); 1925 assign(z64, mkU64(0)); 1951 assign(f_dep1, mkU64(nzc [all...] |
H A D | guest_amd64_toIR.c | 281 static IRExpr* mkU64 ( ULong i ) function 292 case Ity_I64: return mkU64(i); 1670 = mkIRExprVec_5( mkU64(cond), 1799 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); 1825 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); 1827 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); 1869 mkU64(ccOp), 1901 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(ccOp)) ); 1903 stmt( IRStmt_Put( OFFB_CC_DEP2, mkU64(0)) ); 1915 stmt( IRStmt_Put( OFFB_CC_OP, mkU64(base_o [all...] |
H A D | guest_arm_toIR.c | 277 static IRExpr* mkU64 ( ULong i ) function 405 return binop(Iop_64HLtoV128, mkU64(i), mkU64(i)); 2930 assign(old_mask, mkU64(0)); 2931 assign(old_res, mkU64(0)); 2945 assign(new_arg, binop(Iop_Sub8x8, mkexpr(old_arg), mkU64(imm))); 2946 assign(cur_mask, binop(cmp, mkU64(imm), mkexpr(old_arg))); 3089 imm_val = binop(Iop_64HLtoV128, mkU64(imm), mkU64(imm)); 3092 imm_val = mkU64(im [all...] |
H A D | guest_arm64_helpers.c | 719 # define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 946 mkU64(1)); 954 mkU64(1)), 955 mkU64(1)); 1128 # undef mkU64 macro
|
H A D | guest_x86_toIR.c | 680 static IRExpr* mkU64 ( ULong i ) function 5753 mkU64(0) 5814 ? mkU64(0) 7750 assign(one32x2, mkU64( (1ULL << 32) + 1 )); 7814 assign( zero, mkU64(0) ); 7855 assign( zero, mkU64(0) ); 9309 putXMMRegLane64( gregOfRM(modrm), 1, mkU64(0) ); 10590 putXMMRegLane64( gregOfRM(modrm), 1, mkU64(0) ); 10598 putXMMRegLane64( gregOfRM(modrm), 1, mkU64(0) ); 11412 assign( lo64r, mkU64( [all...] |
H A D | host_s390_isel.c | 185 mkU64(ULong value) function 1971 addr_lo = IRExpr_Binop(Iop_Add64, addr_hi, mkU64(8)); 2660 addr_lo = IRExpr_Binop(Iop_Add64, addr_hi, mkU64(8));
|
/external/valgrind/coregrind/ |
H A D | m_translate.c | 991 static IRExpr* mkU64 ( ULong n ) { function 1033 IRExpr*(*mkU)(ULong) = mkU64; 1126 IRExpr*(*mkU)(ULong) = mkU64; 1217 addStmtToIRSB( bb, IRStmt_Put( offB_LR, mkU64( bogus_RA )) ); 1218 addStmtToIRSB( bb, IRStmt_Put( offB_GPR2, mkU64( new_R2_value )) ); 1247 bb->next = IRExpr_Binop(Iop_And64, IRExpr_RdTmp(old_LR), mkU64(~(3ULL))); 1297 addStmtToIRSB( bb, IRStmt_Put( offB_LR, mkU64( bogus_RA )) ); 1332 nraddr_szB == 8 ? mkU64(0) : mkU32(0) 1344 addStmtToIRSB(bb, IRStmt_Put(offB_GPR25, mkU64(closure->readdr))); 1352 VG_WORDSIZE==8 ? mkU64( [all...] |
/external/valgrind/exp-dhat/ |
H A D | dh_main.c | 768 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 790 IRStmt* st2 = assign(t2, binop(Iop_Add64, mkexpr(t1), mkU64(n))); 846 : binop(Iop_Sub64, mkexpr(sp), mkU64(rz_szB))) 864 : binop(Iop_CmpLT64U, mkU64(THRESH), mkexpr(diff))) 1022 #undef mkU64 macro
|
/external/valgrind/memcheck/ |
H A D | mc_translate.c | 444 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 907 = assignNew('V', mce, Ity_I64, binop(Iop_And64, pcdTo64, mkU64(1))); 922 = assignNew('V', mce, Ity_I64, binop(Iop_And64, pcd, mkU64(1))); 1014 top = mkU64(0xFFFFFFFFFFFFFFFFULL); 1130 threeLeft1 = m64 ? mkU64(3<<1) : mkU32(3<<1); 1153 sevenLeft1 = m64 ? mkU64(7<<1) : mkU32(7<<1); 1993 one = mkU64(1); 4708 eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); 5137 eBiasQ0 = tyAddr==Ity_I32 ? mkU32(bias+offQ0) : mkU64(bias+offQ0); 5146 eBiasQ1 = tyAddr==Ity_I32 ? mkU32(bias+offQ1) : mkU64(bia [all...] |
/external/valgrind/VEX/useful/ |
H A D | test_main.c | 741 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 1001 tmp1 = assignNew(mce, Ity_I1, binop(Iop_CmpNE64, vbits, mkU64(0))); 2093 eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); 2309 eBias0 = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias); 2316 eBias8 = tyAddr==Ity_I32 ? mkU32(bias+8) : mkU64(bias+8); 2335 eBias = tyAddr==Ity_I32 ? mkU32(bias) : mkU64(bias);
|
/external/valgrind/helgrind/ |
H A D | hg_main.c | 4396 #define mkU64(_n) IRExpr_Const(IRConst_U64(_n)) macro 4558 : binop(Iop_Add64, mkexpr(addr_minus_sp), mkU64(rz_szB))) 4568 : binop(Iop_CmpLT64U, mkU64(THRESH), mkexpr(diff))) 4867 #undef mkU64 macro
|