/external/valgrind/main/VEX/priv/ |
H A D | guest_s390_helpers.c | 2284 binop(Iop_Or64, cc_dep1, cc_dep2), 2290 binop(Iop_Or64, cc_dep1, cc_dep2), 2319 binop(Iop_Or64, cc_dep1, cc_dep2), 2325 binop(Iop_Or64, cc_dep1, cc_dep2),
|
H A D | guest_arm64_toIR.c | 414 case Ity_I64: return Iop_Or64; 1284 binop(Iop_Or64, IRExpr_Get(OFFB_CC_OP, Ity_I64), cond), 1596 binop(Iop_Or64, 1652 assign(t16, binop(Iop_Or64, mkexpr(src), 1655 assign(t32, binop(Iop_Or64, mkexpr(t16), 1658 assign(t64, binop(Iop_Or64, mkexpr(t32), 1664 assign(t32, binop(Iop_Or64, mkexpr(src), 1667 assign(t64, binop(Iop_Or64, mkexpr(t32), 1673 assign(t64, binop(Iop_Or64, mkexpr(src), 1735 binop(Iop_Or64, [all...] |
H A D | ir_opt.c | 1275 case Iop_Or64: 1685 case Iop_Or64: 2080 case Iop_Or64: 5159 if (is_Binop(aa, Iop_Or64) 5163 IRExpr_Binop(Iop_Or64, 5167 if (is_Binop(aa, Iop_Or64) 5171 IRExpr_Binop(Iop_Or64,
|
H A D | guest_mips_toIR.c | 1563 assign(tmpRd, binop(Iop_Or64, mkexpr(tmpL), mkexpr(tmpR))); 1586 assign(tmpRd, binop(Iop_Or64, mkexpr(tmpL), mkexpr(tmpR))); 1623 assign(tmpOr, binop(Iop_Or64, mkexpr(tmpRd), mkexpr(tmpR))); 14637 assign(tmpT7, binop(Iop_Or64, mkexpr(tmpT2), mkexpr(tmpT4))); 14638 assign(tmpRd, binop(Iop_Or64, mkexpr(tmpT6), mkexpr(tmpT7))); 14684 assign(tmpT9, binop(Iop_Or64, mkexpr(tmpT8), mkexpr(tmpT4))); 14685 assign(tmpRd, binop(Iop_Or64, mkexpr(tmpT2), mkexpr(tmpT9))); 14733 assign(tmpT9, binop(Iop_Or64, mkexpr(tmpT2), mkexpr(tmpT8))); 14734 assign(tmpRd, binop(Iop_Or64, mkexpr(tmpT6), mkexpr(tmpT9))); 14760 assign(tmpRd, binop(Iop_Or64, mkexp [all...] |
H A D | guest_s390_toIR.c | 5843 s390_irgen_load_and_bitwise64(r1, r3, op2addr, Iop_Or64); 6911 assign(result, binop(Iop_Or64, mkexpr(op1), mkexpr(op2))); 6943 assign(result, binop(Iop_Or64, mkexpr(op2), mkexpr(op3))); 6991 assign(result, binop(Iop_Or64, mkexpr(op1), mkexpr(op2))); 7451 put_gpr_dw0(r1, binop(Iop_Or64, binop(Iop_Shl64, mkexpr(op), unop(Iop_64to8, 7474 assign(op2, rot == 0 ? get_gpr_dw0(r2) : binop(Iop_Or64, binop(Iop_Shl64, 7489 put_gpr_dw0(r1, binop(Iop_Or64, binop(Iop_And64, get_gpr_dw0(r1), 7513 assign(op2, rot == 0 ? get_gpr_dw0(r2) : binop(Iop_Or64, binop(Iop_Shl64, 7528 put_gpr_dw0(r1, binop(Iop_Or64, binop(Iop_And64, get_gpr_dw0(r1), 7552 assign(op2, rot == 0 ? get_gpr_dw0(r2) : binop(Iop_Or64, bino [all...] |
H A D | guest_arm_toIR.c | 2947 assign(new_mask, binop(Iop_Or64, mkexpr(old_mask), mkexpr(cur_mask))); 2948 assign(new_res, binop(Iop_Or64, 2964 assign(new_res, binop(Iop_Or64, 3352 assign(res, binop(Iop_Or64, mkexpr(arg_n), 3371 assign(res, binop(Iop_Or64, mkexpr(arg_n), 3410 binop(Iop_Or64, 3437 binop(Iop_Or64, 3464 binop(Iop_Or64, 3803 assign(shval, binop(Q ? Iop_OrV128 : Iop_Or64, 3935 assign(shval, binop(Q ? Iop_OrV128 : Iop_Or64, [all...] |
H A D | host_x86_isel.c | 1905 unop(Iop_CmpNEZ64, binop(Iop_Or64, bind(0), bind(1)))); 2229 case Iop_Or64: 2235 X86AluOp op = e->Iex.Binop.op==Iop_Or64 ? Xalu_OR 2724 unop(Iop_CmpwNEZ64,binop(Iop_Or64,bind(0),bind(1))));
|
H A D | guest_amd64_toIR.c | 3863 assign( t2m, binop(Iop_Or64, mkU64(mask), mkexpr(t2)) ); 5142 put_C3210( binop(Iop_Or64, 7108 case 0xEB: op = Iop_Or64; break; 7681 binop(Iop_Or64, 7723 binop(Iop_Or64, 7744 binop(Iop_Or64, 8282 binop(Iop_Or64, 8308 = binop(Iop_Or64, binop(Iop_And64, mkexpr(flags), mkU64(mask_SZACP)), 8311 = binop(Iop_Or64, rax_with_hole, 9618 binop(Iop_Or64, [all...] |
H A D | guest_x86_toIR.c | 5625 case 0xEB: op = Iop_Or64; break; 6145 binop(Iop_Or64, 7811 binop(Iop_Or64, 7849 binop(Iop_Or64, 7859 binop(Iop_Or64, 11417 binop( Iop_Or64, 11547 binop( Iop_Or64, 12583 binop(Iop_Or64, 12796 Iop_Or64, 12823 Iop_Or64, [all...] |
H A D | guest_ppc_toIR.c | 1531 rot = binop(Iop_Or64, 3103 binop( Iop_Or64, 4310 res = binop( Iop_Or64, 4510 assign( rot, binop(Iop_Or64, r, 4513 binop(Iop_Or64, 4546 assign( rot, binop(Iop_Or64, mkexpr(rTmp), 4593 assign(rot, binop(Iop_Or64, r, binop(Iop_Shl64, r, mkU8(32)))); 4692 assign( rA, binop(Iop_Or64, 5861 IROp opOR = is32bit ? Iop_Or32 : Iop_Or64; 13563 binop(Iop_Or64, mkexp [all...] |
H A D | host_mips_isel.c | 853 case Iop_Or64: 2469 case Iop_Or64: 2475 MIPSAluOp op = (op_binop == Iop_Or64) ? Malu_OR :
|
H A D | host_ppc_isel.c | 1408 case Iop_Or8: case Iop_Or16: case Iop_Or32: case Iop_Or64: 3170 case Iop_Or64: 3176 PPCAluOp op = (op_binop == Iop_Or64) ? Palu_OR :
|
H A D | ir_defs.c | 130 case Iop_Or8 ... Iop_Or64: 2437 case Iop_Or64: case Iop_And64: case Iop_Xor64:
|
H A D | host_amd64_isel.c | 990 case Iop_Or8: case Iop_Or16: case Iop_Or32: case Iop_Or64: 2296 unop(Iop_CmpNEZ64, binop(Iop_Or64, bind(0), bind(1))));
|
H A D | guest_amd64_helpers.c | 1403 Iop_Or64,
|
H A D | host_arm64_isel.c | 1702 case Iop_Or64: case Iop_Or32: lop = ARM64lo_OR; goto log_binop; 2547 //ZZ case Iop_Or64: { 2780 //ZZ case Iop_Or64: {
|
H A D | host_s390_isel.c | 1452 case Iop_Or64:
|
H A D | host_arm_isel.c | 2002 case Iop_Or64: { 2235 case Iop_Or64: {
|
/external/valgrind/main/memcheck/ |
H A D | mc_translate.c | 549 return assignNew('V', mce, Ity_I64, binop(Iop_Or64, a1, a2)); 560 tmp5 = assignNew('V', mce, Ity_I64, binop(Iop_Or64, tmp1, tmp3)); 561 tmp6 = assignNew('V', mce, Ity_I64, binop(Iop_Or64, tmp2, tmp4)); 649 return assignNew('V', mce, Ity_I64, binop(Iop_Or64, data, vbits)); 714 binop(Iop_Or64, 817 IRAtom* tmp4 = assignNew('V', mce, Ity_I64, binop(Iop_Or64, tmp2, tmp3)); 939 opOR = Iop_Or64; 941 opUIFU = Iop_Or64; 1031 IROp opOR = m64 ? Iop_Or64 : Iop_Or32; 1840 opOR = Iop_Or64; [all...] |
/external/valgrind/main/VEX/ |
H A D | test_main.c | 793 return assignNew(mce, Ity_I64, binop(Iop_Or64, a1, a2)); 881 return assignNew(mce, Ity_I64, binop(Iop_Or64, data, vbits)); 938 binop(Iop_Or64, 1899 case Iop_Or64:
|
/external/valgrind/main/VEX/pub/ |
H A D | libvex_ir.h | 424 Iop_Or8, Iop_Or16, Iop_Or32, Iop_Or64, enumerator in enum:__anon32282
|
/external/valgrind/main/memcheck/tests/vbit-test/ |
H A D | util.c | 258 case Iop_Or64: case Iop_And64: case Iop_Xor64:
|
H A D | irops.c | 28 { DEFOP(Iop_Or64, UNDEF_OR), .s390x = 1, .amd64 = 1, .x86 = 1, .arm = 1, .ppc64 = 1, .ppc32 = 1, .mips32 = 0, .mips64 = 1 }, // mips asserts
|