Searched refs:size (Results 1 - 25 of 63) sorted by relevance

123

/kernel/debug/kdb/
H A Dkdb_support.c60 static char *kdb_name_table[100]; /* arbitrary size */
87 #define knt1_size 128 /* must be >= kallsyms table size */
322 * size Size of the area.
326 int kdb_getarea_size(void *res, unsigned long addr, size_t size) argument
328 int ret = probe_kernel_read((char *)res, (char *)addr, size);
347 * size Size of the area.
351 int kdb_putarea_size(unsigned long addr, void *res, size_t size) argument
353 int ret = probe_kernel_read((char *)addr, (char *)res, size);
373 * size Size of the area
377 static int kdb_getphys(void *res, unsigned long addr, size_t size) argument
403 kdb_getphysword(unsigned long *word, unsigned long addr, size_t size) argument
453 kdb_getword(unsigned long *word, unsigned long addr, size_t size) argument
502 kdb_putword(unsigned long addr, unsigned long word, size_t size) argument
694 u32 size; member in struct:debug_alloc_header
745 debug_kmalloc(size_t size, gfp_t flags) argument
[all...]
/kernel/events/
H A Dinternal.h89 unsigned long size, written; \
92 size = min(handle->size, len); \
93 written = memcpy_func(handle->addr, buf, size); \
94 written = size - written; \
99 handle->size -= written; \
100 if (!handle->size) { \
106 handle->size = PAGE_SIZE << page_order(rb); \
108 } while (len && written == size); \
H A Dring_buffer.c105 struct perf_event *event, unsigned int size)
135 size += sizeof(lost_event);
137 size += event->id_header_size;
146 unlikely(CIRC_SPACE(head, tail, perf_data_size(rb)) < size))
161 head += size;
177 handle->size = (1UL << page_shift) - offset;
182 lost_event.header.size = sizeof(lost_event);
279 unsigned long size; local
282 size = sizeof(struct ring_buffer);
283 size
104 perf_output_begin(struct perf_output_handle *handle, struct perf_event *event, unsigned int size) argument
385 unsigned long size; local
[all...]
H A Dcallchain.c62 int size; local
70 size = offsetof(struct callchain_cpus_entries, cpu_entries[nr_cpu_ids]);
72 entries = kzalloc(size, GFP_KERNEL);
76 size = sizeof(struct perf_callchain_entry) * PERF_NR_CONTEXTS;
79 entries->cpu_entries[cpu] = kmalloc_node(size, GFP_KERNEL,
H A Dcore.c1183 int size = 0; local
1187 size += sizeof(u64);
1190 size += sizeof(u64);
1197 size += sizeof(u64);
1200 size += entry * nr;
1201 event->read_size = size;
1208 u16 size = 0; local
1213 size += sizeof(data->ip);
1216 size += sizeof(data->addr);
1219 size
1240 u16 size = 0; local
3550 int n = 0, size = 0, ret = -EFAULT; local
4161 unsigned long size = perf_data_size(rb); local
4541 u64 size = 0; local
4779 int size = 1; local
4800 u32 size; member in struct:__anon21
4812 size_t size; local
4894 int size = 1; local
4905 int size = sizeof(u32); local
4917 int size = sizeof(u64); /* nr */ local
4927 int size = sizeof(u64); local
4948 u16 size = sizeof(u64); local
5120 int ret, size = task_event->event_id.header.size; local
5213 int size = comm_event->event_id.header.size; local
5243 unsigned int size; local
5326 int size = mmap_event->event_id.header.size; local
5379 unsigned int size; local
7038 u32 size; local
[all...]
/kernel/bpf/
H A Dtest_stub.c44 int size; member in struct:bpf_context_access
57 static bool test_is_valid_access(int off, int size, enum bpf_access_type type) argument
65 if (access->size == size && (access->type & type))
H A Dcore.c57 void *bpf_internal_load_pointer_neg_helper(const struct sk_buff *skb, int k, unsigned int size) argument
65 if (ptr >= skb->head && ptr + size <= skb_tail_pointer(skb))
71 struct bpf_prog *bpf_prog_alloc(unsigned int size, gfp_t gfp_extra_flags) argument
78 size = round_up(size, PAGE_SIZE);
79 fp = __vmalloc(size, gfp_flags, PAGE_KERNEL);
89 fp->pages = size / PAGE_SIZE;
96 struct bpf_prog *bpf_prog_realloc(struct bpf_prog *fp_old, unsigned int size, argument
105 size = round_up(size, PAGE_SIZ
139 unsigned int size, hole, start; local
[all...]
/kernel/sched/
H A Dcpudeadline.c59 if ((l < cp->size) && dl_time_before(cp->elements[idx].dl,
62 if ((r < cp->size) && dl_time_before(cp->elements[largest].dl,
155 new_cpu = cp->elements[cp->size - 1].cpu;
156 cp->elements[old_idx].dl = cp->elements[cp->size - 1].dl;
158 cp->size--;
174 cp->size++;
175 cp->elements[cp->size - 1].dl = 0;
176 cp->elements[cp->size - 1].cpu = cpu;
177 cp->elements[cpu].idx = cp->size - 1;
178 cpudl_change_key(cp, cp->size
[all...]
H A Dcpudeadline.h16 int size; member in struct:cpudl
/kernel/
H A Dtaskstats.c80 size_t size)
88 skb = genlmsg_new(size, GFP_KERNEL);
431 size_t size; local
444 size = nla_total_size(sizeof(struct cgroupstats));
447 size);
509 size_t size; local
511 size = nla_total_size(sizeof(u32)) +
514 size += nla_total_size(0); /* Padding for alignment */
516 return size;
523 size_t size; local
79 prepare_reply(struct genl_info *info, u8 cmd, struct sk_buff **skbp, size_t size) argument
552 size_t size; local
622 size_t size; local
[all...]
H A Dresource.c257 resource_size_t size; local
270 /* need to restore size, and keep flags */
271 size = resource_size(tmp);
273 tmp->end = size - 1;
502 int region_is_ram(resource_size_t start, unsigned long size) argument
505 resource_size_t end = start + size - 1;
536 resource_size_t size,
557 resource_size_t size,
591 size, constraint->align);
592 alloc.end = alloc.start + size
534 simple_align_resource(void *data, const struct resource *avail, resource_size_t size, resource_size_t align) argument
555 __find_resource(struct resource *root, struct resource *old, struct resource *new, resource_size_t size, struct resource_constraint *constraint) argument
613 find_resource(struct resource *root, struct resource *new, resource_size_t size, struct resource_constraint *constraint) argument
681 allocate_resource(struct resource *root, struct resource *new, resource_size_t size, resource_size_t min, resource_size_t max, resource_size_t align, resource_size_t (*alignf)(void *, const struct resource *, resource_size_t, resource_size_t), void *alignf_data) argument
864 __adjust_resource(struct resource *res, resource_size_t start, resource_size_t size) argument
911 adjust_resource(struct resource *res, resource_size_t start, resource_size_t size) argument
1199 release_mem_region_adjustable(struct resource *parent, resource_size_t start, resource_size_t size) argument
1449 iomem_map_sanity_check(resource_size_t addr, unsigned long size) argument
1504 int size = PAGE_SIZE; local
[all...]
/kernel/printk/
H A Dprintk.c284 /* Return log buffer size */
376 /* compute the message size including the padding bytes */
379 u32 size; local
381 size = sizeof(struct printk_log) + text_len + dict_len;
382 *pad_len = (-size) & (LOG_ALIGN - 1);
383 size += *pad_len;
385 return size;
410 /* compute the size again, count also the warning message */
421 u32 size, pad_len; local
425 size
839 log_buf_len_update(unsigned size) argument
850 unsigned size = memparse(str, &str); local
1045 msg_print_text(const struct printk_log *msg, enum log_flags prev, bool syslog, char *buf, size_t size) argument
1104 syslog_print(char __user *buf, int size) argument
1168 syslog_print_all(char __user *buf, int size, bool clear) argument
1592 cont_print_text(char *text, size_t size) argument
1883 msg_print_text(const struct printk_log *msg, enum log_flags prev, bool syslog, char *buf, size_t size) argument
1885 cont_print_text(char *text, size_t size) argument
2132 console_cont_flush(char *text, size_t size) argument
2811 kmsg_dump_get_line_nolock(struct kmsg_dumper *dumper, bool syslog, char *line, size_t size, size_t *len) argument
2860 kmsg_dump_get_line(struct kmsg_dumper *dumper, bool syslog, char *line, size_t size, size_t *len) argument
2893 kmsg_dump_get_buffer(struct kmsg_dumper *dumper, bool syslog, char *buf, size_t size, size_t *len) argument
[all...]
/kernel/gcov/
H A Dbase.c117 static inline int within(void *addr, void *start, unsigned long size) argument
119 return ((addr >= start) && (addr < start + size));
H A Dgcc_3_4.c214 /* Get size of function info entry. Based on gcc magic. */
217 size_t size; local
219 size = sizeof(struct gcov_fn_info) + num_counter_active(info) *
222 size = ALIGN(size, __alignof__(struct gcov_fn_info));
223 return size;
267 size_t size = ctr->num * sizeof(gcov_type); local
271 dup->counts[i].values = vmalloc(size);
274 memcpy(dup->counts[i].values, ctr->values, size);
H A Dgcc_4_7.c258 size_t fi_size; /* function info size */
259 size_t cv_size; /* counter values size */
355 * @size: size of buffer
361 size_t size; member in struct:gcov_iterator
483 /* Dry-run to get the actual buffer size. */
484 iter->size = convert_to_gcda(NULL, info);
485 iter->buffer = vmalloc(iter->size);
535 if (iter->pos < iter->size)
538 if (iter->pos >= iter->size)
[all...]
/kernel/trace/
H A Dtrace_stack.c35 * us to remove most or all of the stack size overhead
56 int size; local
67 size = stack_dump_index[i];
69 size = stack_dump_index[i] - stack_dump_index[i+1];
72 size, (void *)stack_dump_trace[i]);
353 int size; local
375 size = stack_dump_index[i];
377 size = stack_dump_index[i] - stack_dump_index[i+1];
379 seq_printf(m, "%3ld) %8d %5d ", i, stack_dump_index[i], size);
H A Dtrace_probe.c282 return NULL; /* Only string type needs size function */
498 bprm->hi_shift = BYTES_TO_BITS(t->size) - (bw + bo);
501 return (BYTES_TO_BITS(t->size) < (bw + bo)) ? -EINVAL : 0;
505 int traceprobe_parse_probe_arg(char *arg, ssize_t *size, argument
534 parg->offset = *size;
535 *size += parg->type->size;
619 size_t size; local
626 size = count - done;
628 if (size >
[all...]
H A Dtrace_syscalls.c313 int size; local
331 size = sizeof(*entry) + sizeof(unsigned long) * sys_data->nb_args;
338 sys_data->enter_event->event.type, size, irq_flags, pc);
567 int size; local
583 /* get the size after alignment with the u32 buffer size field */
584 size = sizeof(unsigned long) * sys_data->nb_args + sizeof(*rec);
585 size = ALIGN(size + sizeof(u32), sizeof(u64));
586 size
641 int size; local
[all...]
H A Dtrace_uprobe.c537 ret = traceprobe_parse_probe_arg(arg, &tu->tp.size, parg,
779 int size, esize; local
784 if (WARN_ON_ONCE(tu->tp.size + dsize > PAGE_SIZE))
791 size = esize + tu->tp.size + dsize;
793 call->event.type, size, 0, 0);
807 memcpy(data, ucb->buf, tu->tp.size + dsize);
975 int ret, i, size; local
982 size = SIZEOF_TRACE_ENTRY(true);
985 size
1103 int size, esize; local
[all...]
H A Dtrace_kprobe.c762 ret = traceprobe_parse_probe_arg(arg, &tk->tp.size, parg,
924 int size, dsize, pc; local
937 size = sizeof(*entry) + tk->tp.size + dsize;
941 size, irq_flags, pc);
972 int size, pc, dsize; local
985 size = sizeof(*entry) + tk->tp.size + dsize;
989 size, irq_flags, pc);
1107 parg->type->size,
1149 int size, __size, dsize; local
1180 int size, __size, dsize; local
[all...]
H A Dtrace_export.c42 #define __array(type, item, size) type item[size];
45 #define __array_desc(type, container, item, size) type item[size];
H A Dtrace.c47 * On boot up, the ring buffer is set to the minimum size, so that
318 * trace_buf_size is the size in bytes that is allocated
320 * to page size.
460 * @size: The size of the string.
462 int __trace_puts(unsigned long ip, const char *str, int size) argument
479 alloc = sizeof(*entry) + size + 2; /* possible \n added */
491 memcpy(&entry->buf, str, size);
494 if (entry->buf[size - 1] != '\n') {
495 entry->buf[size]
518 int size = sizeof(struct bputs_entry); local
836 trace_parser_get_init(struct trace_parser *parser, int size) argument
1773 int size = FTRACE_STACK_ENTRIES; local
2108 int len = 0, size, pc; local
2161 int len = 0, size, pc; local
4046 __tracing_resize_ring_buffer(struct trace_array *tr, unsigned long size, int cpu) argument
4112 tracing_resize_ring_buffer(struct trace_array *tr, unsigned long size, int cpu_id) argument
4797 unsigned long size; local
4865 unsigned long size = 0, expanded_size = 0; local
4927 int size; local
5423 ssize_t size; local
5586 int entries, size, i; local
5777 ftrace_arch_read_dyn_info(char *buf, int size) argument
5790 int size = ARRAY_SIZE(ftrace_dyn_info_buffer); local
6319 allocate_trace_buffer(struct trace_array *tr, struct trace_buffer *buf, int size) argument
6344 allocate_trace_buffers(struct trace_array *tr, int size) argument
[all...]
/kernel/power/
H A Duser.c202 loff_t size; local
298 size = snapshot_get_image_size();
299 size <<= PAGE_SHIFT;
300 error = put_user(size, (loff_t __user *)arg);
304 size = count_swap_pages(data->swap, 1);
305 size <<= PAGE_SHIFT;
306 error = put_user(size, (loff_t __user *)arg);
H A Dsnapshot.c57 * Preferred image size in bytes (tunable via /sys/power/image_size).
59 * size will not exceed N bytes, but if that is impossible, it will
187 unsigned int used_space; /* total size of objects allocated out
203 static void *chain_alloc(struct chain_allocator *ca, unsigned int size) argument
207 if (LINKED_PAGE_DATA_SIZE - ca->used_space < size) {
219 ca->used_space += size;
1523 * minimum_image_size - Estimate the minimum acceptable size of an image
1527 * minimum acceptable size of a hibernation image to use as the lower limit for
1530 * We assume that the minimum image size should be proportional to
1540 unsigned long size; local
1577 unsigned long saveable, size, max_size, count, highmem, pages = 0; local
[all...]
/kernel/irq/
H A Dirqdomain.c29 * @size: Size of linear map; 0 for radix mapping only
39 struct irq_domain *__irq_domain_add(struct device_node *of_node, int size, argument
46 domain = kzalloc_node(sizeof(*domain) + (sizeof(unsigned int) * size),
57 domain->revmap_size = size;
108 * @size: total number of irqs in mapping
124 unsigned int size,
131 domain = __irq_domain_add(of_node, size, size, 0, ops, host_data);
138 int rc = irq_alloc_descs(first_irq, first_irq, size,
144 irq_domain_associate_many(domain, first_irq, 0, size);
123 irq_domain_add_simple(struct device_node *of_node, unsigned int size, unsigned int first_irq, const struct irq_domain_ops *ops, void *host_data) argument
166 irq_domain_add_legacy(struct device_node *of_node, unsigned int size, unsigned int first_irq, irq_hw_number_t first_hwirq, const struct irq_domain_ops *ops, void *host_data) argument
[all...]

Completed in 239 milliseconds

123