Searched refs:insn (Results 1 - 5 of 5) sorted by relevance
/kernel/bpf/ |
H A D | verifier.c | 23 * All paths of conditional branches are analyzed until 'bpf_exit' insn. 31 * The second pass is all possible path descent from the 1st insn. 33 * analysis is limited to 32k insn, which may be hit even if total number of 34 * insn is less then 4K, but there are too many branches that change stack/regs. 54 * 1st insn copies R10 (which has FRAME_PTR) type into R1 57 * So after 2nd insn, the register R1 has type PTR_TO_STACK 103 * BPF_MOV64_REG(BPF_REG_2, BPF_REG_10), // after this insn R2 type is FRAME_PTR 104 * BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), // after this insn R2 type is PTR_TO_STACK 105 * BPF_LD_MAP_FD(BPF_REG_1, map_fd), // after this insn R1 type is CONST_PTR_TO_MAP 114 * If it's ok, then verifier allows this BPF_CALL insn an 318 print_bpf_insn(struct bpf_insn *insn) argument 695 check_xadd(struct verifier_env *env, struct bpf_insn *insn) argument 917 check_alu_op(struct reg_state *regs, struct bpf_insn *insn) argument 1045 check_cond_jmp_op(struct verifier_env *env, struct bpf_insn *insn, int *insn_idx) argument 1145 ld_imm64_to_map_ptr(struct bpf_insn *insn) argument 1153 check_ld_imm(struct verifier_env *env, struct bpf_insn *insn) argument 1481 struct bpf_insn *insn; local 1695 struct bpf_insn *insn = env->prog->insnsi; local 1782 struct bpf_insn *insn = env->prog->insnsi; local [all...] |
H A D | core.c | 46 #define DST regs[insn->dst_reg] 47 #define SRC regs[insn->src_reg] 51 #define IMM insn->imm 182 * @insn: is the array of eBPF instructions 186 static unsigned int __bpf_prog_run(void *ctx, const struct bpf_insn *insn) argument 292 #define CONT ({ insn++; goto select_insn; }) 293 #define CONT_JMP ({ insn++; goto select_insn; }) 303 goto *jumptable[insn->code]; 348 DST = (u64) (u32) insn[0].imm | ((u64) (u32) insn[ [all...] |
H A D | syscall.c | 360 /* fixup insn->imm field of bpf_call instructions: 361 * if (insn->imm == BPF_FUNC_map_lookup_elem) 362 * insn->imm = bpf_map_lookup_elem - __bpf_call_base; 363 * else if (insn->imm == BPF_FUNC_map_update_elem) 364 * insn->imm = bpf_map_update_elem - __bpf_call_base; 375 struct bpf_insn *insn = &prog->insnsi[i]; local 377 if (insn->code == (BPF_JMP | BPF_CALL)) { 384 fn = prog->aux->ops->get_func_proto(insn->imm); 389 insn->imm = fn->func - __bpf_call_base;
|
/kernel/events/ |
H A D | uprobes.c | 80 * insn - copy_insn() saves the original instruction here for 218 * @insn: instruction to be checked. 220 * Returns true if @insn is a breakpoint instruction. 222 bool __weak is_swbp_insn(uprobe_opcode_t *insn) argument 224 return *insn == UPROBE_SWBP_INSN; 229 * @insn: instruction to be checked. 231 * Returns true if @insn is a breakpoint instruction. 236 bool __weak is_trap_insn(uprobe_opcode_t *insn) argument 238 return is_swbp_insn(insn); 366 return uprobe_write_opcode(mm, vaddr, *(uprobe_opcode_t *)&auprobe->insn); 537 __copy_insn(struct address_space *mapping, struct file *filp, void *insn, int nbytes, loff_t offset) argument 563 void *insn = &uprobe->arch.insn; local 1167 uprobe_opcode_t insn = UPROBE_SWBP_INSN; local [all...] |
/kernel/ |
H A D | signal.c | 1150 unsigned char insn; local 1152 if (get_user(insn, (unsigned char *)(regs->ip + i))) 1154 printk(KERN_CONT "%02x ", insn);
|
Completed in 44 milliseconds