/mm/ |
H A D | quicklist.c | 28 unsigned long node_free_pages, max; local 42 max = node_free_pages / FRACTION_OF_NODE_MEM; 45 max /= num_cpus_on_node; 47 return max(max, min_pages);
|
H A D | readahead.c | 248 * for 128k (32 page) max ra 251 static unsigned long get_init_ra_size(unsigned long size, unsigned long max) argument 255 if (newsize <= max / 32) 257 else if (newsize <= max / 4) 260 newsize = max; 270 unsigned long max) 275 if (cur < max / 16) 280 return min(newsize, max); 323 * Count contiguously cached pages from @offset-1 to @offset-@max, 329 pgoff_t offset, unsigned long max) 269 get_next_ra_size(struct file_ra_state *ra, unsigned long max) argument 328 count_history_pages(struct address_space *mapping, pgoff_t offset, unsigned long max) argument 343 try_context_readahead(struct address_space *mapping, struct file_ra_state *ra, pgoff_t offset, unsigned long req_size, unsigned long max) argument 383 unsigned long max = max_sane_readahead(ra->ra_pages); local [all...] |
H A D | bootmem.c | 359 unsigned long max; local 367 max = min(bdata->node_low_pfn, end); 369 err = mark_bootmem_node(bdata, pos, max, reserve, flags); 375 if (max == end) 497 unsigned long min, max, start, sidx, midx, step; local 511 max = bdata->node_low_pfn; 516 if (limit && max > limit) 517 max = limit; 518 if (max <= min) 521 step = max(alig [all...] |
H A D | memblock.c | 37 .memory.max = INIT_MEMBLOCK_REGIONS, 41 .reserved.max = INIT_MEMBLOCK_REGIONS, 46 .physmem.max = INIT_PHYSMEM_REGIONS, 203 end = max(start, end); 214 bottom_up_start = max(start, kernel_end); 288 memblock.reserved.max); 300 memblock.memory.max); 337 old_size = type->max * sizeof(struct memblock_region); 383 memblock_type_name(type), type->max, type->max * [all...] |
H A D | cma.c | 171 alignment = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order); 241 alignment = max(alignment, 242 (phys_addr_t)PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order));
|
H A D | page_cgroup.c | 362 unsigned long idx, max; local 374 max = idx; 375 for (idx = 0; idx < max; idx++)
|
H A D | page_io.c | 163 page_no < sis->max) { 217 sis->max = page_no;
|
H A D | util.c | 43 * @max: read at most @max chars from @s 46 char *kstrndup(const char *s, size_t max, gfp_t gfp) argument 54 len = strnlen(s, max);
|
H A D | page_alloc.c | 573 * We restrict max order of merging to prevent merge 3950 return max(size, 4UL); 3957 * The max wait table size = 4096 x sizeof(wait_queue_head_t). ie: 4255 pageset_update(&p->pcp, 6 * batch, max(1UL, 1 * batch)); 4284 unsigned long batch = max(1UL, high / 4); 4509 * @nid: The nid to return the range for. If MAX_NUMNODES, the min and max PFN are returned. 4529 *end_pfn = max(*end_pfn, this_end_pfn); 4618 zone_start_pfn = max(zone_start_pfn, node_start_pfn); 5180 required_kernelcore = max(required_kernelcore, corepages); 5215 start_pfn = max(start_pf 5598 long max = 0; local 5981 unsigned long long max = high_limit; local [all...] |
H A D | page-writeback.c | 47 #define MAX_PAUSE max(HZ/5, 1) 58 #define BANDWIDTH_INTERVAL max(HZ/5, 1) 515 return max(thresh, global_dirty_limit); 805 bdi_thresh = max(bdi_thresh, (limit - dirty) / 8); 910 thresh = max(thresh, dirty); 1105 bdi->dirty_ratelimit = max(dirty_ratelimit, 1UL); 1219 t = max(1, HZ / 100);
|
H A D | swapfile.c | 470 while (tmp < si->max && tmp < (cluster_next(&cluster->index) + 1) * 587 si->lowest_bit = si->max; 739 if (offset >= p->max) 1285 unsigned int max = si->max; local 1296 if (++i >= max) { 1305 max = prev + 1; 1725 ret = add_swap_extent(sis, 0, sis->max, 0); 1734 ret = add_swap_extent(sis, 0, sis->max, 0); 1906 p->max [all...] |
H A D | percpu.c | 107 int contig_hint; /* max contiguous size hint */ 208 return max(highbit - PCPU_SLOT_BASE_SHIFT + 2, 1); 582 max_contig = max(this_size, max_contig); 597 max_contig = max(*p - p[-1], max_contig); 625 max_contig = max(head, max_contig); 629 max_contig = max(tail, max_contig); 640 chunk->contig_hint = max(chunk->contig_hint, 720 chunk->contig_hint = max(chunk->map[i + 1] - chunk->map[i] - 1, chunk->contig_hint); 863 * @align: alignment of area (max PAGE_SIZE) 1057 * @align: alignment of area (max PAGE_SIZ [all...] |
H A D | frontswap.c | 324 bitmap_zero(sis->frontswap_map, sis->max);
|
H A D | slub.c | 851 slab_err(s, page, "objects %u > max %u", 856 slab_err(s, page, "inuse %u > max %u", 2746 for (order = max(min_order, 3047 if (oo_objects(s->oo) > oo_objects(s->max)) 3048 s->max = s->oo; 3360 int objects = oo_objects(s->max); 3633 s->object_size = max(s->object_size, (int)size); 3848 unsigned long *map = kmalloc(BITS_TO_LONGS(oo_objects(s->max)) * 3879 unsigned long max; member in struct:loc_track 3886 if (t->max) 3891 alloc_loc_track(struct loc_track *t, unsigned long max, gfp_t flags) argument [all...] |
H A D | mmap.c | 360 unsigned long max, subtree_gap; local 361 max = vma->vm_start; 363 max -= vma->vm_prev->vm_end; 367 if (subtree_gap > max) 368 max = subtree_gap; 373 if (subtree_gap > max) 374 max = subtree_gap; 376 return max; 598 max(addr, vma->vm_start)) >> PAGE_SHIFT; 1985 info.low_limit = max(PAGE_SIZ [all...] |
H A D | process_vm_access.c | 171 nr_pages = max(nr_pages, nr_pages_iov);
|
H A D | memory.c | 203 batch->max = MAX_GATHER_BATCH; 228 tlb->local.max = ARRAY_SIZE(tlb->__pages); 299 if (batch->nr == batch->max) { 304 VM_BUG_ON_PAGE(batch->nr > batch->max, page); 306 return batch->max - batch->nr; 1314 unsigned long start = max(vma->vm_start, start_addr); 2849 start_addr = max(address & mask, vma->vm_start);
|
H A D | slab.c | 154 #define REDZONE_ALIGN max(BYTES_PER_WORD, __alignof__(unsigned long long)) 812 struct array_cache *from, unsigned int max) 815 int nr = min3(from->avail, max, to->limit - to->avail); 3308 int max = shared_array->limit - shared_array->avail; local 3309 if (max) { 3310 if (batchcount > max) 3311 batchcount = max; 811 transfer_objects(struct array_cache *to, struct array_cache *from, unsigned int max) argument
|
H A D | slab_common.c | 289 align = max(align, ralign);
|
H A D | memory_hotplug.c | 299 zone->spanned_pages = max(old_zone_end_pfn, end_pfn) - 439 pgdat->node_spanned_pages = max(old_pgdat_end_pfn, end_pfn) -
|
H A D | vmscan.c | 2813 .nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX), 2989 sc->nr_to_reclaim = max(SWAP_CLUSTER_MAX, high_wmark_pages(zone)); 3649 .nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX),
|
H A D | compaction.c | 1330 rc = max(status, rc);
|
H A D | vmstat.c | 103 threshold = max(1, (int)(watermark_distance / num_online_cpus()));
|
H A D | kmemleak.c | 178 /* max number of lines to be printed */ 567 max_addr = max(max_addr, ptr + size);
|
H A D | vmalloc.c | 503 vmap_area_pcpu_hole = max(vmap_area_pcpu_hole, va->va_end); 594 * *end = max(*end, highest purged address) 745 #define VMAP_MAX(x, y) ((x) > (y) ? (x) : (y)) /* can't use max() */
|