/mm/ |
H A D | quicklist.c | 28 unsigned long node_free_pages, max; local 42 max = node_free_pages / FRACTION_OF_NODE_MEM; 45 max /= num_cpus_on_node; 47 return max(max, min_pages);
|
H A D | readahead.c | 265 * for 128k (32 page) max ra 268 static unsigned long get_init_ra_size(unsigned long size, unsigned long max) argument 272 if (newsize <= max / 32) 274 else if (newsize <= max / 4) 277 newsize = max; 287 unsigned long max) 292 if (cur < max / 16) 297 return min(newsize, max); 340 * Count contiguously cached pages from @offset-1 to @offset-@max, 347 pgoff_t offset, unsigned long max) 286 get_next_ra_size(struct file_ra_state *ra, unsigned long max) argument 345 count_history_pages(struct address_space *mapping, struct file_ra_state *ra, pgoff_t offset, unsigned long max) argument 361 try_context_readahead(struct address_space *mapping, struct file_ra_state *ra, pgoff_t offset, unsigned long req_size, unsigned long max) argument 401 unsigned long max = max_sane_readahead(ra->ra_pages); local [all...] |
H A D | util.c | 38 * @max: read at most @max chars from @s 41 char *kstrndup(const char *s, size_t max, gfp_t gfp) argument 49 len = strnlen(s, max);
|
H A D | page_cgroup.c | 357 unsigned long idx, max; local 369 max = idx; 370 for (idx = 0; idx < max; idx++)
|
H A D | bootmem.c | 331 unsigned long max; local 339 max = min(bdata->node_low_pfn, end); 341 err = mark_bootmem_node(bdata, pos, max, reserve, flags); 347 if (max == end) 475 unsigned long min, max, start, sidx, midx, step; local 489 max = bdata->node_low_pfn; 494 if (limit && max > limit) 495 max = limit; 496 if (max <= min) 499 step = max(alig [all...] |
H A D | swapfile.c | 220 si->lowest_alloc = si->max; 311 si->lowest_bit = si->max; 489 if (offset >= p->max) 1027 unsigned int max = si->max; local 1038 if (++i >= max) { 1047 max = prev + 1; 1440 ret = add_swap_extent(sis, 0, sis->max, 0); 1456 page_no < sis->max) { 1510 sis->max [all...] |
H A D | slab.c | 148 #define REDZONE_ALIGN max(BYTES_PER_WORD, __alignof__(unsigned long long)) 961 struct array_cache *from, unsigned int max) 964 int nr = min3(from->avail, max, to->limit - to->avail); 3633 int max = shared_array->limit - shared_array->avail; local 3634 if (max) { 3635 if (batchcount > max) 3636 batchcount = max; 960 transfer_objects(struct array_cache *to, struct array_cache *from, unsigned int max) argument
|
H A D | page_alloc.c | 3426 return max(size, 4UL); 3433 * The max wait table size = 4096 x sizeof(wait_queue_head_t). ie: 3699 pcp->batch = max(1UL, 1 * batch); 3716 pcp->batch = max(1UL, high/4); 3960 * @nid: The nid to return the range for. If MAX_NUMNODES, the min and max PFN are returned. 3980 *end_pfn = max(*end_pfn, this_end_pfn); 4069 zone_start_pfn = max(zone_start_pfn, node_start_pfn); 4554 required_kernelcore = max(required_kernelcore, corepages); 4590 start_pfn = max(start_pfn, zone_movable_pfn[nid]); 4681 * @max_zone_pfn: an array of max PFN 4867 unsigned long max = 0; local 5204 unsigned long long max = limit; local [all...] |
H A D | slub.c | 844 slab_err(s, page, "objects %u > max %u", 849 slab_err(s, page, "inuse %u > max %u", 2644 for (order = max(min_order, 2730 align = max(align, ralign); 2981 if (oo_objects(s->oo) > oo_objects(s->max)) 2982 s->max = s->oo; 3468 int objects = oo_objects(s->max); 3912 s->objsize = max(s->objsize, (int)size); 4130 unsigned long *map = kmalloc(BITS_TO_LONGS(oo_objects(s->max)) * 4163 unsigned long max; member in struct:loc_track 4175 alloc_loc_track(struct loc_track *t, unsigned long max, gfp_t flags) argument [all...] |