/mm/ |
H A D | internal.h | 204 * the maximally aligned gigantic page 'base'. Handle any discontiguity 207 static inline struct page *mem_map_offset(struct page *base, int offset) argument 210 return pfn_to_page(page_to_pfn(base) + offset); 211 return base + offset; 216 * page 'base'. Handle any discontiguity in the mem_map. 219 struct page *base, int offset) 222 unsigned long pfn = page_to_pfn(base) + offset; 218 mem_map_next(struct page *iter, struct page *base, int offset) argument
|
H A D | page_cgroup.c | 35 struct page_cgroup *base; local 37 base = NODE_DATA(page_to_nid(page))->node_page_cgroup; 38 if (unlikely(!base)) 42 return base + offset; 60 struct page_cgroup *base, *pc; local 72 base = __alloc_bootmem_node_nopanic(NODE_DATA(nid), 74 if (!base) 77 pc = base + index; 80 NODE_DATA(nid)->node_page_cgroup = base; 167 struct page_cgroup *base, *p local 211 struct page_cgroup *base; local [all...] |
H A D | slob.c | 200 slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK); local 201 slobidx_t offset = next - base; 225 slob_t *base = (slob_t *)((unsigned long)s & PAGE_MASK); local 232 return base+next;
|
H A D | bootmem.c | 452 unsigned long base = bdata->node_min_pfn; local 459 return ALIGN(base + idx, step) - base; 465 unsigned long base = PFN_PHYS(bdata->node_min_pfn); local 469 return ALIGN(base + off, align) - base;
|
H A D | memblock.c | 61 long __init_memblock memblock_overlaps_region(struct memblock_type *type, phys_addr_t base, phys_addr_t size) argument 66 phys_addr_t rgnbase = type->regions[i].base; 68 if (memblock_addrs_overlap(base, size, rgnbase, rgnsize)) 83 phys_addr_t base, res_base; local 90 base = memblock_align_down((end - size), align); 98 while (start <= base) { 99 j = memblock_overlaps_region(&memblock.reserved, base, size); 101 return base; 102 res_base = memblock.reserved.regions[j].base; 105 base 276 memblock_add_region(struct memblock_type *type, phys_addr_t base, phys_addr_t size) argument 406 memblock_add(phys_addr_t base, phys_addr_t size) argument 412 __memblock_remove(struct memblock_type *type, phys_addr_t base, phys_addr_t size) argument 463 memblock_remove(phys_addr_t base, phys_addr_t size) argument 468 memblock_free(phys_addr_t base, phys_addr_t size) argument 473 memblock_reserve(phys_addr_t base, phys_addr_t size) argument 705 memblock_is_region_memory(phys_addr_t base, phys_addr_t size) argument 716 memblock_is_region_reserved(phys_addr_t base, phys_addr_t size) argument 729 unsigned long long base, size; local [all...] |
H A D | filemap.c | 1994 const struct iovec *iov, size_t base, size_t bytes) 1999 char __user *buf = iov->iov_base + base; 2000 int copy = min(bytes, iov->iov_len - base); 2002 base = 0; 2079 size_t base = i->iov_offset; local 2088 copy = min(bytes, iov->iov_len - base); 2092 base += copy; 2093 if (iov->iov_len == base) { 2095 base = 0; 2099 i->iov_offset = base; 1993 __iovec_copy_from_user_inatomic(char *vaddr, const struct iovec *iov, size_t base, size_t bytes) argument [all...] |
H A D | nommu.c | 1130 void *base; local 1187 base = page_address(pages); 1189 region->vm_start = (unsigned long) base; 1206 ret = vma->vm_file->f_op->read(vma->vm_file, base, len, &fpos); 1214 memset(base + ret, 0, len - ret);
|
H A D | percpu.c | 27 * Percpu access can be done by configuring percpu base registers 103 void *base_addr; /* base address of this chunk */ 803 /* return address relative to base address */ 957 void __percpu *base = __addr_to_pcpu_ptr(pcpu_base_addr); local 961 void *start = per_cpu_ptr(base, cpu); 985 void __percpu *base = __addr_to_pcpu_ptr(pcpu_base_addr); local 1001 void *start = per_cpu_ptr(base, cpu); 1594 void *base = (void *)ULONG_MAX; local 1614 /* allocate, copy and determine base address */ 1632 base [all...] |
H A D | swapfile.c | 2308 pgoff_t base, end; local 2316 base = (target >> our_page_cluster) << our_page_cluster; 2317 end = base + (1 << our_page_cluster); 2318 if (!base) /* first page is swap header */ 2319 base++; 2334 for (toff = target; --toff >= base; nr_pages++) {
|
H A D | vmalloc.c | 1407 * @addr: base address 1488 * @addr: memory base address 1508 * @addr: memory base address 2315 * existing vmap_area, the base address is pulled down to fit the 2328 unsigned long base, start, end, last_end; local 2383 base = vmalloc_end - last_end; 2386 base = pvm_determine_end(&next, &prev, align) - end; 2389 BUG_ON(next && next->va_end <= base + end); 2390 BUG_ON(prev && prev->va_end > base + end); 2393 * base migh [all...] |
H A D | slub.c | 245 void *base; local 250 base = page_address(page); 251 if (object < base || object >= base + page->objects * s->size || 252 (object - base) % s->size) {
|