/kernel/debug/kdb/ |
H A D | kdb_support.c | 60 static char *kdb_name_table[100]; /* arbitrary size */ 87 #define knt1_size 128 /* must be >= kallsyms table size */ 322 * size Size of the area. 326 int kdb_getarea_size(void *res, unsigned long addr, size_t size) argument 328 int ret = probe_kernel_read((char *)res, (char *)addr, size); 347 * size Size of the area. 351 int kdb_putarea_size(unsigned long addr, void *res, size_t size) argument 353 int ret = probe_kernel_read((char *)addr, (char *)res, size); 373 * size Size of the area 377 static int kdb_getphys(void *res, unsigned long addr, size_t size) argument 403 kdb_getphysword(unsigned long *word, unsigned long addr, size_t size) argument 453 kdb_getword(unsigned long *word, unsigned long addr, size_t size) argument 502 kdb_putword(unsigned long addr, unsigned long word, size_t size) argument 694 u32 size; member in struct:debug_alloc_header 745 debug_kmalloc(size_t size, gfp_t flags) argument [all...] |
/kernel/events/ |
H A D | internal.h | 89 unsigned long size, written; \ 92 size = min(handle->size, len); \ 93 written = memcpy_func(handle->addr, buf, size); \ 94 written = size - written; \ 99 handle->size -= written; \ 100 if (!handle->size) { \ 106 handle->size = PAGE_SIZE << page_order(rb); \ 108 } while (len && written == size); \
|
H A D | ring_buffer.c | 105 struct perf_event *event, unsigned int size) 135 size += sizeof(lost_event); 137 size += event->id_header_size; 146 unlikely(CIRC_SPACE(head, tail, perf_data_size(rb)) < size)) 161 head += size; 177 handle->size = (1UL << page_shift) - offset; 182 lost_event.header.size = sizeof(lost_event); 279 unsigned long size; local 282 size = sizeof(struct ring_buffer); 283 size 104 perf_output_begin(struct perf_output_handle *handle, struct perf_event *event, unsigned int size) argument 385 unsigned long size; local [all...] |
H A D | callchain.c | 62 int size; local 70 size = offsetof(struct callchain_cpus_entries, cpu_entries[nr_cpu_ids]); 72 entries = kzalloc(size, GFP_KERNEL); 76 size = sizeof(struct perf_callchain_entry) * PERF_NR_CONTEXTS; 79 entries->cpu_entries[cpu] = kmalloc_node(size, GFP_KERNEL,
|
H A D | core.c | 1183 int size = 0; local 1187 size += sizeof(u64); 1190 size += sizeof(u64); 1197 size += sizeof(u64); 1200 size += entry * nr; 1201 event->read_size = size; 1208 u16 size = 0; local 1213 size += sizeof(data->ip); 1216 size += sizeof(data->addr); 1219 size 1240 u16 size = 0; local 3550 int n = 0, size = 0, ret = -EFAULT; local 4161 unsigned long size = perf_data_size(rb); local 4541 u64 size = 0; local 4779 int size = 1; local 4800 u32 size; member in struct:__anon21 4812 size_t size; local 4894 int size = 1; local 4905 int size = sizeof(u32); local 4917 int size = sizeof(u64); /* nr */ local 4927 int size = sizeof(u64); local 4948 u16 size = sizeof(u64); local 5120 int ret, size = task_event->event_id.header.size; local 5213 int size = comm_event->event_id.header.size; local 5243 unsigned int size; local 5326 int size = mmap_event->event_id.header.size; local 5379 unsigned int size; local 7038 u32 size; local [all...] |
/kernel/bpf/ |
H A D | test_stub.c | 44 int size; member in struct:bpf_context_access 57 static bool test_is_valid_access(int off, int size, enum bpf_access_type type) argument 65 if (access->size == size && (access->type & type))
|
H A D | core.c | 57 void *bpf_internal_load_pointer_neg_helper(const struct sk_buff *skb, int k, unsigned int size) argument 65 if (ptr >= skb->head && ptr + size <= skb_tail_pointer(skb)) 71 struct bpf_prog *bpf_prog_alloc(unsigned int size, gfp_t gfp_extra_flags) argument 78 size = round_up(size, PAGE_SIZE); 79 fp = __vmalloc(size, gfp_flags, PAGE_KERNEL); 89 fp->pages = size / PAGE_SIZE; 96 struct bpf_prog *bpf_prog_realloc(struct bpf_prog *fp_old, unsigned int size, argument 105 size = round_up(size, PAGE_SIZ 139 unsigned int size, hole, start; local [all...] |
/kernel/sched/ |
H A D | cpudeadline.c | 59 if ((l < cp->size) && dl_time_before(cp->elements[idx].dl, 62 if ((r < cp->size) && dl_time_before(cp->elements[largest].dl, 155 new_cpu = cp->elements[cp->size - 1].cpu; 156 cp->elements[old_idx].dl = cp->elements[cp->size - 1].dl; 158 cp->size--; 174 cp->size++; 175 cp->elements[cp->size - 1].dl = 0; 176 cp->elements[cp->size - 1].cpu = cpu; 177 cp->elements[cpu].idx = cp->size - 1; 178 cpudl_change_key(cp, cp->size [all...] |
H A D | cpudeadline.h | 16 int size; member in struct:cpudl
|
/kernel/ |
H A D | taskstats.c | 80 size_t size) 88 skb = genlmsg_new(size, GFP_KERNEL); 431 size_t size; local 444 size = nla_total_size(sizeof(struct cgroupstats)); 447 size); 509 size_t size; local 511 size = nla_total_size(sizeof(u32)) + 514 size += nla_total_size(0); /* Padding for alignment */ 516 return size; 523 size_t size; local 79 prepare_reply(struct genl_info *info, u8 cmd, struct sk_buff **skbp, size_t size) argument 552 size_t size; local 622 size_t size; local [all...] |
H A D | resource.c | 257 resource_size_t size; local 270 /* need to restore size, and keep flags */ 271 size = resource_size(tmp); 273 tmp->end = size - 1; 502 int region_is_ram(resource_size_t start, unsigned long size) argument 505 resource_size_t end = start + size - 1; 536 resource_size_t size, 557 resource_size_t size, 591 size, constraint->align); 592 alloc.end = alloc.start + size 534 simple_align_resource(void *data, const struct resource *avail, resource_size_t size, resource_size_t align) argument 555 __find_resource(struct resource *root, struct resource *old, struct resource *new, resource_size_t size, struct resource_constraint *constraint) argument 613 find_resource(struct resource *root, struct resource *new, resource_size_t size, struct resource_constraint *constraint) argument 681 allocate_resource(struct resource *root, struct resource *new, resource_size_t size, resource_size_t min, resource_size_t max, resource_size_t align, resource_size_t (*alignf)(void *, const struct resource *, resource_size_t, resource_size_t), void *alignf_data) argument 864 __adjust_resource(struct resource *res, resource_size_t start, resource_size_t size) argument 911 adjust_resource(struct resource *res, resource_size_t start, resource_size_t size) argument 1199 release_mem_region_adjustable(struct resource *parent, resource_size_t start, resource_size_t size) argument 1449 iomem_map_sanity_check(resource_size_t addr, unsigned long size) argument 1504 int size = PAGE_SIZE; local [all...] |
/kernel/printk/ |
H A D | printk.c | 284 /* Return log buffer size */ 376 /* compute the message size including the padding bytes */ 379 u32 size; local 381 size = sizeof(struct printk_log) + text_len + dict_len; 382 *pad_len = (-size) & (LOG_ALIGN - 1); 383 size += *pad_len; 385 return size; 410 /* compute the size again, count also the warning message */ 421 u32 size, pad_len; local 425 size 839 log_buf_len_update(unsigned size) argument 850 unsigned size = memparse(str, &str); local 1045 msg_print_text(const struct printk_log *msg, enum log_flags prev, bool syslog, char *buf, size_t size) argument 1104 syslog_print(char __user *buf, int size) argument 1168 syslog_print_all(char __user *buf, int size, bool clear) argument 1592 cont_print_text(char *text, size_t size) argument 1883 msg_print_text(const struct printk_log *msg, enum log_flags prev, bool syslog, char *buf, size_t size) argument 1885 cont_print_text(char *text, size_t size) argument 2132 console_cont_flush(char *text, size_t size) argument 2811 kmsg_dump_get_line_nolock(struct kmsg_dumper *dumper, bool syslog, char *line, size_t size, size_t *len) argument 2860 kmsg_dump_get_line(struct kmsg_dumper *dumper, bool syslog, char *line, size_t size, size_t *len) argument 2893 kmsg_dump_get_buffer(struct kmsg_dumper *dumper, bool syslog, char *buf, size_t size, size_t *len) argument [all...] |
/kernel/gcov/ |
H A D | base.c | 117 static inline int within(void *addr, void *start, unsigned long size) argument 119 return ((addr >= start) && (addr < start + size));
|
H A D | gcc_3_4.c | 214 /* Get size of function info entry. Based on gcc magic. */ 217 size_t size; local 219 size = sizeof(struct gcov_fn_info) + num_counter_active(info) * 222 size = ALIGN(size, __alignof__(struct gcov_fn_info)); 223 return size; 267 size_t size = ctr->num * sizeof(gcov_type); local 271 dup->counts[i].values = vmalloc(size); 274 memcpy(dup->counts[i].values, ctr->values, size);
|
H A D | gcc_4_7.c | 258 size_t fi_size; /* function info size */ 259 size_t cv_size; /* counter values size */ 355 * @size: size of buffer 361 size_t size; member in struct:gcov_iterator 483 /* Dry-run to get the actual buffer size. */ 484 iter->size = convert_to_gcda(NULL, info); 485 iter->buffer = vmalloc(iter->size); 535 if (iter->pos < iter->size) 538 if (iter->pos >= iter->size) [all...] |
/kernel/trace/ |
H A D | trace_stack.c | 35 * us to remove most or all of the stack size overhead 56 int size; local 67 size = stack_dump_index[i]; 69 size = stack_dump_index[i] - stack_dump_index[i+1]; 72 size, (void *)stack_dump_trace[i]); 353 int size; local 375 size = stack_dump_index[i]; 377 size = stack_dump_index[i] - stack_dump_index[i+1]; 379 seq_printf(m, "%3ld) %8d %5d ", i, stack_dump_index[i], size);
|
H A D | trace_probe.c | 282 return NULL; /* Only string type needs size function */ 498 bprm->hi_shift = BYTES_TO_BITS(t->size) - (bw + bo); 501 return (BYTES_TO_BITS(t->size) < (bw + bo)) ? -EINVAL : 0; 505 int traceprobe_parse_probe_arg(char *arg, ssize_t *size, argument 534 parg->offset = *size; 535 *size += parg->type->size; 619 size_t size; local 626 size = count - done; 628 if (size > [all...] |
H A D | trace_syscalls.c | 313 int size; local 331 size = sizeof(*entry) + sizeof(unsigned long) * sys_data->nb_args; 338 sys_data->enter_event->event.type, size, irq_flags, pc); 567 int size; local 583 /* get the size after alignment with the u32 buffer size field */ 584 size = sizeof(unsigned long) * sys_data->nb_args + sizeof(*rec); 585 size = ALIGN(size + sizeof(u32), sizeof(u64)); 586 size 641 int size; local [all...] |
H A D | trace_uprobe.c | 537 ret = traceprobe_parse_probe_arg(arg, &tu->tp.size, parg, 779 int size, esize; local 784 if (WARN_ON_ONCE(tu->tp.size + dsize > PAGE_SIZE)) 791 size = esize + tu->tp.size + dsize; 793 call->event.type, size, 0, 0); 807 memcpy(data, ucb->buf, tu->tp.size + dsize); 975 int ret, i, size; local 982 size = SIZEOF_TRACE_ENTRY(true); 985 size 1103 int size, esize; local [all...] |
H A D | trace_kprobe.c | 762 ret = traceprobe_parse_probe_arg(arg, &tk->tp.size, parg, 924 int size, dsize, pc; local 937 size = sizeof(*entry) + tk->tp.size + dsize; 941 size, irq_flags, pc); 972 int size, pc, dsize; local 985 size = sizeof(*entry) + tk->tp.size + dsize; 989 size, irq_flags, pc); 1107 parg->type->size, 1149 int size, __size, dsize; local 1180 int size, __size, dsize; local [all...] |
H A D | trace_export.c | 42 #define __array(type, item, size) type item[size]; 45 #define __array_desc(type, container, item, size) type item[size];
|
H A D | trace.c | 47 * On boot up, the ring buffer is set to the minimum size, so that 318 * trace_buf_size is the size in bytes that is allocated 320 * to page size. 460 * @size: The size of the string. 462 int __trace_puts(unsigned long ip, const char *str, int size) argument 479 alloc = sizeof(*entry) + size + 2; /* possible \n added */ 491 memcpy(&entry->buf, str, size); 494 if (entry->buf[size - 1] != '\n') { 495 entry->buf[size] 518 int size = sizeof(struct bputs_entry); local 836 trace_parser_get_init(struct trace_parser *parser, int size) argument 1773 int size = FTRACE_STACK_ENTRIES; local 2108 int len = 0, size, pc; local 2161 int len = 0, size, pc; local 4046 __tracing_resize_ring_buffer(struct trace_array *tr, unsigned long size, int cpu) argument 4112 tracing_resize_ring_buffer(struct trace_array *tr, unsigned long size, int cpu_id) argument 4797 unsigned long size; local 4865 unsigned long size = 0, expanded_size = 0; local 4927 int size; local 5423 ssize_t size; local 5586 int entries, size, i; local 5777 ftrace_arch_read_dyn_info(char *buf, int size) argument 5790 int size = ARRAY_SIZE(ftrace_dyn_info_buffer); local 6319 allocate_trace_buffer(struct trace_array *tr, struct trace_buffer *buf, int size) argument 6344 allocate_trace_buffers(struct trace_array *tr, int size) argument [all...] |
/kernel/power/ |
H A D | user.c | 202 loff_t size; local 298 size = snapshot_get_image_size(); 299 size <<= PAGE_SHIFT; 300 error = put_user(size, (loff_t __user *)arg); 304 size = count_swap_pages(data->swap, 1); 305 size <<= PAGE_SHIFT; 306 error = put_user(size, (loff_t __user *)arg);
|
H A D | snapshot.c | 57 * Preferred image size in bytes (tunable via /sys/power/image_size). 59 * size will not exceed N bytes, but if that is impossible, it will 187 unsigned int used_space; /* total size of objects allocated out 203 static void *chain_alloc(struct chain_allocator *ca, unsigned int size) argument 207 if (LINKED_PAGE_DATA_SIZE - ca->used_space < size) { 219 ca->used_space += size; 1523 * minimum_image_size - Estimate the minimum acceptable size of an image 1527 * minimum acceptable size of a hibernation image to use as the lower limit for 1530 * We assume that the minimum image size should be proportional to 1540 unsigned long size; local 1577 unsigned long saveable, size, max_size, count, highmem, pages = 0; local [all...] |
/kernel/irq/ |
H A D | irqdomain.c | 29 * @size: Size of linear map; 0 for radix mapping only 39 struct irq_domain *__irq_domain_add(struct device_node *of_node, int size, argument 46 domain = kzalloc_node(sizeof(*domain) + (sizeof(unsigned int) * size), 57 domain->revmap_size = size; 108 * @size: total number of irqs in mapping 124 unsigned int size, 131 domain = __irq_domain_add(of_node, size, size, 0, ops, host_data); 138 int rc = irq_alloc_descs(first_irq, first_irq, size, 144 irq_domain_associate_many(domain, first_irq, 0, size); 123 irq_domain_add_simple(struct device_node *of_node, unsigned int size, unsigned int first_irq, const struct irq_domain_ops *ops, void *host_data) argument 166 irq_domain_add_legacy(struct device_node *of_node, unsigned int size, unsigned int first_irq, irq_hw_number_t first_hwirq, const struct irq_domain_ops *ops, void *host_data) argument [all...] |