/arch/x86/mm/kmemcheck/ |
H A D | opcode.c | 39 void kmemcheck_opcode_decode(const uint8_t *op, unsigned int *size) argument 45 for (; opcode_is_prefix(*op); ++op) { 46 if (*op == 0x66) 51 if (opcode_is_rex_prefix(*op)) { 52 uint8_t rex = *op; 54 ++op; 56 switch (*op) { 61 ++op; 63 switch (*op) { 98 kmemcheck_opcode_get_primary(const uint8_t *op) argument [all...] |
H A D | opcode.h | 6 void kmemcheck_opcode_decode(const uint8_t *op, unsigned int *size); 7 const uint8_t *kmemcheck_opcode_get_primary(const uint8_t *op);
|
H A D | selftest.c | 38 static bool selftest_opcode_one(const struct selftest_opcode *op) argument 42 kmemcheck_opcode_decode(op->insn, &size); 44 if (size == op->expected_size) 48 op->desc, op->expected_size, size);
|
/arch/sh/include/asm/ |
H A D | atomic-llsc.h | 18 #define ATOMIC_OP(op) \ 19 static inline void atomic_##op(int i, atomic_t *v) \ 24 "1: movli.l @%2, %0 ! atomic_" #op "\n" \ 25 " " #op " %1, %0 \n" \ 33 #define ATOMIC_OP_RETURN(op) \ 34 static inline int atomic_##op##_return(int i, atomic_t *v) \ 39 "1: movli.l @%2, %0 ! atomic_" #op "_return \n" \ 40 " " #op " %1, %0 \n" \ 51 #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETUR [all...] |
H A D | atomic-grb.h | 4 #define ATOMIC_OP(op) \ 5 static inline void atomic_##op(int i, atomic_t *v) \ 15 " " #op " %2, %0 \n\t" /* $op */ \ 24 #define ATOMIC_OP_RETURN(op) \ 25 static inline int atomic_##op##_return(int i, atomic_t *v) \ 35 " " #op " %2, %0 \n\t" /* $op */ \ 46 #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETUR [all...] |
H A D | atomic-irq.h | 12 #define ATOMIC_OP(op, c_op) \ 13 static inline void atomic_##op(int i, atomic_t *v) \ 22 #define ATOMIC_OP_RETURN(op, c_op) \ 23 static inline int atomic_##op##_return(int i, atomic_t *v) \ 36 #define ATOMIC_OPS(op, c_op) ATOMIC_OP(op, c_op) ATOMIC_OP_RETURN(op, c_op)
|
/arch/ia64/include/asm/sn/ |
H A D | mspec.h | 26 #define FETCHOP_LOAD_OP(addr, op) ( \ 27 *(volatile long *)((char*) (addr) + (op))) 29 #define FETCHOP_STORE_OP(addr, op, x) ( \ 30 *(volatile long *)((char*) (addr) + (op)) = (long) (x))
|
/arch/sh/kernel/ |
H A D | kgdb.c | 21 #define OPCODE_BT(op) (((op) & 0xff00) == 0x8900) 22 #define OPCODE_BF(op) (((op) & 0xff00) == 0x8b00) 23 #define OPCODE_BTF_DISP(op) (((op) & 0x80) ? (((op) | 0xffffff80) << 1) : \ 24 (((op) & 0x7f ) << 1)) 25 #define OPCODE_BFS(op) (((op) 50 insn_size_t op = __raw_readw(linux_regs->pc); local [all...] |
/arch/x86/include/asm/ |
H A D | rmwcc.h | 16 #define GEN_UNARY_RMWcc(op, var, arg0, cc) \ 17 __GEN_RMWcc(op " " arg0, var, cc) 19 #define GEN_BINARY_RMWcc(op, var, vcon, val, arg0, cc) \ 20 __GEN_RMWcc(op " %1, " arg0, var, cc, vcon (val)) 33 #define GEN_UNARY_RMWcc(op, var, arg0, cc) \ 34 __GEN_RMWcc(op " " arg0, var, cc) 36 #define GEN_BINARY_RMWcc(op, var, vcon, val, arg0, cc) \ 37 __GEN_RMWcc(op " %2, " arg0, var, cc, vcon (val))
|
H A D | paravirt_types.h | 366 #define paravirt_type(op) \ 367 [paravirt_typenum] "i" (PARAVIRT_PATCH(op)), \ 368 [paravirt_opptr] "i" (&(op)) 534 #define PVOP_TEST_NULL(op) BUG_ON(op == NULL) 536 #define PVOP_TEST_NULL(op) ((void)op) 539 #define ____PVOP_CALL(rettype, op, clbr, call_clbr, extra_clbr, \ 544 PVOP_TEST_NULL(op); \ 552 : paravirt_type(op), \ [all...] |
/arch/arm/include/asm/ |
H A D | firmware.h | 50 * call_firmware_op(op, ...) 55 #define call_firmware_op(op, ...) \ 56 ((firmware_ops->op) ? firmware_ops->op(__VA_ARGS__) : (-ENOSYS))
|
/arch/x86/kernel/kprobes/ |
H A D | opt.c | 44 struct optimized_kprobe *op; local 53 op = container_of(kp, struct optimized_kprobe, kp); 54 /* If op->list is not empty, op is under optimizing */ 55 if (list_empty(&op->list)) 65 * bytes must be recovered from op->optinsn.copied_insn buffer. 70 memcpy(buf + 1, op->optinsn.copied_insn, RELATIVE_ADDR_SIZE); 73 memcpy(buf, op->optinsn.copied_insn + offs, RELATIVE_ADDR_SIZE - offs); 142 optimized_callback(struct optimized_kprobe *op, struct pt_regs *regs) argument 147 /* This is possible if op i 280 arch_check_optimized_kprobe(struct optimized_kprobe *op) argument 295 arch_within_optimized_kprobe(struct optimized_kprobe *op, unsigned long addr) argument 304 __arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty) argument 313 arch_remove_optimized_kprobe(struct optimized_kprobe *op) argument 323 arch_prepare_optimized_kprobe(struct optimized_kprobe *op) argument 381 struct optimized_kprobe *op, *tmp; local 405 arch_unoptimize_kprobe(struct optimized_kprobe *op) argument 423 struct optimized_kprobe *op, *tmp; local 433 struct optimized_kprobe *op; local [all...] |
/arch/sparc/lib/ |
H A D | atomic_64.S | 18 #define ATOMIC_OP(op) \ 19 ENTRY(atomic_##op) /* %o0 = increment, %o1 = atomic_ptr */ \ 22 op %g1, %o0, %g7; \ 30 ENDPROC(atomic_##op); \ 32 #define ATOMIC_OP_RETURN(op) \ 33 ENTRY(atomic_##op##_return) /* %o0 = increment, %o1 = atomic_ptr */ \ 36 op %g1, %o0, %g7; \ 40 op %g1, %o0, %g1; \ 44 ENDPROC(atomic_##op##_return); 46 #define ATOMIC_OPS(op) ATOMIC_O [all...] |
/arch/ia64/include/asm/ |
H A D | paravirt_privop.h | 240 #define PARAVIRT_OP(op) \ 241 [op_addr] "i"(&pv_cpu_ops.op) 299 #define PARAVIRT_BR0(op, type) \ 304 : PARAVIRT_OP(op) \ 307 #define PARAVIRT_BR0_RET(op, type) \ 312 : PARAVIRT_OP(op) \ 315 #define PARAVIRT_BR1(op, type, arg1) \ 321 : PARAVIRT_OP(op), "0"(__##arg1) \ 324 #define PARAVIRT_BR1_RET(op, type, arg1) \ 330 : PARAVIRT_OP(op), " [all...] |
/arch/sparc/kernel/ |
H A D | of_device_32.c | 244 static void __init build_device_resources(struct platform_device *op, argument 258 bus->count_cells(op->dev.of_node, &na, &ns); 260 preg = of_get_property(op->dev.of_node, bus->addr_prop_name, &num_reg); 270 op->resource = op->archdata.resource; 271 op->num_resources = num_reg; 273 struct resource *r = &op->resource[index]; 276 struct device_node *dp = op->dev.of_node; 326 op->dev.of_node->full_name, index, 334 r->name = op 341 struct platform_device *op = kzalloc(sizeof(*op), GFP_KERNEL); local 396 struct platform_device *op = scan_one_device(dp, parent); local [all...] |
H A D | of_device_common.c | 17 struct platform_device *op = of_find_device_by_node(node); local 19 if (!op || index >= op->archdata.num_irqs) 22 return op->archdata.irqs[index]; 29 struct platform_device *op = of_find_device_by_node(node); local 31 if (!op || index >= op->num_resources) 34 memcpy(r, &op->archdata.resource[index], sizeof(*r)); 41 struct platform_device *op = of_find_device_by_node(node); local 44 if (!op || inde 63 struct platform_device *op = of_find_device_by_node(dp); local [all...] |
H A D | central.c | 63 static int clock_board_probe(struct platform_device *op) argument 73 p->clock_freq_regs = of_ioremap(&op->resource[0], 0, 74 resource_size(&op->resource[0]), 81 p->clock_regs = of_ioremap(&op->resource[1], 0, 82 resource_size(&op->resource[1]), 89 if (op->resource[2].flags) { 90 p->clock_ver_reg = of_ioremap(&op->resource[2], 0, 91 resource_size(&op->resource[2]), 110 p->leds_pdev.dev.parent = &op->dev; 128 of_iounmap(&op 160 fhc_probe(struct platform_device *op) argument [all...] |
/arch/hexagon/include/asm/ |
H A D | atomic.h | 97 #define ATOMIC_OP(op) \ 98 static inline void atomic_##op(int i, atomic_t *v) \ 104 " %0 = "#op "(%0,%2);\n" \ 113 #define ATOMIC_OP_RETURN(op) \ 114 static inline int atomic_##op##_return(int i, atomic_t *v) \ 120 " %0 = "#op "(%0,%2);\n" \ 130 #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op)
|
/arch/metag/include/asm/ |
H A D | atomic_lnkget.h | 30 #define ATOMIC_OP(op) \ 31 static inline void atomic_##op(int i, atomic_t *v) \ 37 " " #op " %0, %0, %2\n" \ 48 #define ATOMIC_OP_RETURN(op) \ 49 static inline int atomic_##op##_return(int i, atomic_t *v) \ 57 " " #op " %1, %1, %3\n" \ 72 #define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op)
|
/arch/powerpc/lib/ |
H A D | sstep.c | 487 #define __put_user_asmx(x, addr, err, op, cr) \ 489 "1: " op " %2,0,%3\n" \ 503 #define __get_user_asmx(x, addr, err, op) \ 505 "1: "op" %1,0,%2\n" \ 518 #define __cacheop_user_asmx(addr, err, op) \ 520 "1: "op" 0,%1\n" \ 640 * Sets *op to indicate what the instruction does. 642 int __kprobes analyse_instr(struct instruction_op *op, struct pt_regs *regs, argument 651 op->type = COMPUTE; 656 op 1769 struct instruction_op op; local [all...] |
/arch/powerpc/kernel/ |
H A D | ftrace.c | 34 unsigned int op; local 38 /* if (link) set op to 'bl' else 'b' */ 39 op = create_branch((unsigned int *)ip, addr, link ? 1 : 0); 41 return op; 87 static int is_bl_op(unsigned int op) argument 89 return (op & 0xfc000003) == 0x48000001; 92 static unsigned long find_bl_target(unsigned long ip, unsigned int op) argument 96 offset = (op & 0x03fffffc); 109 unsigned int op; local 115 if (probe_kernel_read(&op, (voi 174 unsigned int op; local 293 unsigned int op[2]; local 337 unsigned int op; local [all...] |
/arch/mips/mm/ |
H A D | uasm.c | 135 #define I_u1u2u3(op) \ 136 Ip_u1u2u3(op) \ 138 build_insn(buf, insn##op, a, b, c); \ 140 UASM_EXPORT_SYMBOL(uasm_i##op); 142 #define I_s3s1s2(op) \ 143 Ip_s3s1s2(op) \ 145 build_insn(buf, insn##op, b, c, a); \ 147 UASM_EXPORT_SYMBOL(uasm_i##op); 149 #define I_u2u1u3(op) \ 150 Ip_u2u1u3(op) \ [all...] |
/arch/xtensa/include/asm/ |
H A D | atomic.h | 62 #define ATOMIC_OP(op) \ 63 static inline void atomic_##op(int i, atomic_t * v) \ 71 " " #op " %0, %1, %2\n" \ 80 #define ATOMIC_OP_RETURN(op) \ 81 static inline int atomic_##op##_return(int i, atomic_t * v) \ 89 " " #op " %0, %1, %2\n" \ 92 " " #op " %0, %0, %2\n" \ 103 #define ATOMIC_OP(op) \ 104 static inline void atomic_##op(int i, atomic_t * v) \ 111 " " #op " [all...] |
/arch/arc/include/asm/ |
H A D | atomic.h | 26 #define ATOMIC_OP(op, c_op, asm_op) \ 27 static inline void atomic_##op(int i, atomic_t *v) \ 41 #define ATOMIC_OP_RETURN(op, c_op, asm_op) \ 42 static inline int atomic_##op##_return(int i, atomic_t *v) \ 92 #define ATOMIC_OP(op, c_op, asm_op) \ 93 static inline void atomic_##op(int i, atomic_t *v) \ 102 #define ATOMIC_OP_RETURN(op, c_op) \ 103 static inline int atomic_##op##_return(int i, atomic_t *v) \ 119 #define ATOMIC_OPS(op, c_op, asm_op) \ 120 ATOMIC_OP(op, c_o [all...] |
/arch/alpha/include/asm/ |
H A D | atomic.h | 32 #define ATOMIC_OP(op) \ 33 static __inline__ void atomic_##op(int i, atomic_t * v) \ 38 " " #op "l %0,%2,%0\n" \ 48 #define ATOMIC_OP_RETURN(op) \ 49 static inline int atomic_##op##_return(int i, atomic_t *v) \ 55 " " #op "l %0,%3,%2\n" \ 56 " " #op "l %0,%3,%0\n" \ 68 #define ATOMIC64_OP(op) \ 69 static __inline__ void atomic64_##op(long i, atomic64_t * v) \ 74 " " #op " [all...] |