/mm/ |
H A D | nobootmem.c | 36 u64 goal, u64 limit) 41 if (limit > memblock.current_limit) 42 limit = memblock.current_limit; 44 addr = memblock_find_in_range_node(size, align, goal, limit, nid); 224 unsigned long limit) 233 ptr = __alloc_memory_core_early(NUMA_NO_NODE, size, align, goal, limit); 262 unsigned long limit = -1UL; local 264 return ___alloc_bootmem_nopanic(size, align, goal, limit); 268 unsigned long goal, unsigned long limit) 270 void *mem = ___alloc_bootmem_nopanic(size, align, goal, limit); 35 __alloc_memory_core_early(int nid, u64 size, u64 align, u64 goal, u64 limit) argument 221 ___alloc_bootmem_nopanic(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 267 ___alloc_bootmem(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 298 unsigned long limit = -1UL; local 303 ___alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 339 ___alloc_bootmem_node(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument [all...] |
H A D | bootmem.c | 494 unsigned long goal, unsigned long limit) 499 bdebug("nid=%td size=%lx [%lu pages] align=%lx goal=%lx limit=%lx\n", 501 align, goal, limit); 505 BUG_ON(limit && goal + size > limit); 514 limit >>= PAGE_SHIFT; 516 if (limit && max > limit) 517 max = limit; 602 unsigned long limit) 492 alloc_bootmem_bdata(struct bootmem_data *bdata, unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 599 alloc_bootmem_core(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 624 ___alloc_bootmem_nopanic(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 659 unsigned long limit = 0; local 664 ___alloc_bootmem(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 695 unsigned long limit = 0; local 700 ___alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument 739 ___alloc_bootmem_node(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) argument [all...] |
H A D | cma.c | 198 * @limit: End address of the reserved memory (optional, 0 for any). 210 * reserve in range from @base to @limit. 213 phys_addr_t size, phys_addr_t limit, 221 pr_debug("%s(size %pa, base %pa, limit %pa alignment %pa)\n", 222 __func__, &size, &base, &limit, &alignment); 245 limit &= ~(alignment - 1); 266 * If the limit is unspecified or above the memblock end, its effective 270 if (limit == 0 || limit > memblock_end) 271 limit 212 cma_declare_contiguous(phys_addr_t base, phys_addr_t size, phys_addr_t limit, phys_addr_t alignment, unsigned int order_per_bit, bool fixed, struct cma **res_cma) argument [all...] |
H A D | slab.c | 185 * The limit is stored in the per-cpu structure to reduce the data cache 191 unsigned int limit; member in struct:array_cache 462 .limit = BOOT_CPUCACHE_ENTRIES, 660 static void init_arraycache(struct array_cache *ac, int limit, int batch) argument 672 ac->limit = limit; 815 int nr = min3(from->avail, max, to->limit - to->avail); 834 int limit, gfp_t gfp) 877 static struct alien_cache **alloc_alien_cache(int node, int limit, gfp_t gfp) argument 883 if (limit > 833 alloc_alien_cache(int node, int limit, gfp_t gfp) argument 1701 dump_line(char *data, int offset, int limit) argument 1754 int limit; local 1776 int limit; local 3661 __do_tune_cpucache(struct kmem_cache *cachep, int limit, int batchcount, int shared, gfp_t gfp) argument 3702 do_tune_cpucache(struct kmem_cache *cachep, int limit, int batchcount, int shared, gfp_t gfp) argument 3732 int limit = 0; local 4005 int limit, batchcount, shared, res; local [all...] |
H A D | page-writeback.c | 170 * Because the user is allowed to specify the dirty limit globally as 171 * absolute number of bytes, calculating the per-zone dirty limit can 172 * require translating the configured limit into a percentage of 321 * dirty limit, %false if the limit is exceeded. 325 unsigned long limit = zone_dirty_limit(zone); local 329 zone_page_state(zone, NR_WRITEBACK) <= limit; 521 * @dirty: global dirty limit in pages 523 * Returns @bdi's dirty limit in pages. The term "dirty" in the context of 526 * Note that balance_dirty_pages() will only seriously take it as a hard limit 575 pos_ratio_polynom(unsigned long setpoint, unsigned long dirty, unsigned long limit) argument 676 unsigned long limit = hard_dirty_limit(thresh); local 895 unsigned long limit = global_dirty_limit; local 957 unsigned long limit = hard_dirty_limit(thresh); local [all...] |
H A D | sparse.c | 267 unsigned long goal, limit; local 281 limit = goal + (1UL << PA_SECTION_SHIFT); 285 SMP_CACHE_BYTES, goal, limit, 287 if (!p && limit) { 288 limit = 0;
|
H A D | memblock.c | 223 * we always limit bottom-up allocation above the kernel, 224 * but top-down allocation doesn't have the limit, so 1119 * The @min_addr limit is dropped if it can not be satisfied and the allocation 1352 void __init memblock_enforce_memory_limit(phys_addr_t limit) argument 1357 if (!limit) 1362 if (limit <= r->size) { 1363 max_addr = r->base + limit; 1366 limit -= r->size; 1484 void __init_memblock memblock_set_current_limit(phys_addr_t limit) argument 1486 memblock.current_limit = limit; [all...] |
H A D | slab.h | 155 unsigned int limit; member in struct:slabinfo
|
H A D | filemap.c | 2271 unsigned long limit = rlimit(RLIMIT_FSIZE); local 2281 if (limit != RLIM_INFINITY) { 2282 if (*pos >= limit) { 2286 if (*count > limit - (typeof(limit))*pos) { 2287 *count = limit - (typeof(limit))*pos; 2306 * Are we about to exceed the fs block limit ?
|
H A D | memcontrol.c | 169 /* the soft limit is exceeded*/ 456 * limit reclaim to prevent infinite loops, if they ever occur. 729 * limit by the same amount 1007 /* threshold event is triggered in finer grain than soft limit */ 1620 * @memcg: The memory cgroup that went over limit 1641 pr_info(" killed as a result of limit of "); 1647 pr_info("memory: usage %llukB, limit %llukB, failcnt %llu\n", 1651 pr_info("memory+swap: usage %llukB, limit %llukB, failcnt %llu\n", 1655 pr_info("kmem: usage %llukB, limit %llukB, failcnt %llu\n", 1696 * Return the memory (and swap, if configured) limit fo 1700 u64 limit; local 4070 __memcg_activate_kmem(struct mem_cgroup *memcg, unsigned long long limit) argument 4133 memcg_activate_kmem(struct mem_cgroup *memcg, unsigned long long limit) argument 4410 unsigned long long limit, memsw_limit; local [all...] |
H A D | nommu.c | 1670 static int limit; local 1671 if (limit < 5) { 1677 limit++;
|
H A D | mmap.c | 317 * segment grow beyond its set limit the in case where the limit is 1550 /* Check against address space limit. */ 1954 * stack's low limit (the base): 2111 /* address space limit tests */ 2115 /* Stack limit test */ 2119 /* mlock limit tests */ 2122 unsigned long limit; local 2124 limit = ACCESS_ONCE(rlim[RLIMIT_MEMLOCK].rlim_cur); 2125 limit >> [all...] |
H A D | slab_common.c | 827 seq_puts(m, " : tunables <limit> <batchcount> <sharedfactor>"); 898 sinfo.limit, sinfo.batchcount, sinfo.shared);
|