/mm/ |
H A D | quicklist.c | 63 void quicklist_trim(int nr, void (*dtor)(void *), argument 69 q = &get_cpu_var(quicklist)[nr]; 78 void *p = quicklist_alloc(nr, 0, NULL);
|
H A D | filemap_xip.c | 77 unsigned long nr, left; local 82 /* nr is the maximum number of bytes to copy from this page */ 83 nr = PAGE_CACHE_SIZE; 87 nr = ((isize - 1) & ~PAGE_CACHE_MASK) + 1; 88 if (nr <= offset) { 92 nr = nr - offset; 93 if (nr > len - copied) 94 nr = len - copied; 123 left = __copy_to_user(buf+copied, xip_mem+offset, nr); [all...] |
H A D | swap_state.c | 238 void free_pages_and_swap_cache(struct page **pages, int nr) argument 243 while (nr) { 244 int todo = min(nr, PAGEVEC_SIZE); 251 nr -= todo;
|
H A D | vmalloc.c | 91 unsigned long end, pgprot_t prot, struct page **pages, int *nr) 96 * nr is a running index into the array which helps higher level 104 struct page *page = pages[*nr]; 111 (*nr)++; 117 unsigned long end, pgprot_t prot, struct page **pages, int *nr) 127 if (vmap_pte_range(pmd, addr, next, prot, pages, nr)) 134 unsigned long end, pgprot_t prot, struct page **pages, int *nr) 144 if (vmap_pmd_range(pud, addr, next, prot, pages, nr)) 163 int nr = 0; local 169 err = vmap_pud_range(pgd, addr, next, prot, pages, &nr); 90 vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) argument 116 vmap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) argument 133 vmap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) argument 585 int nr = 0; local 2542 unsigned int nr, *counters = m->private; local [all...] |
H A D | highmem.c | 262 unsigned long nr; local 269 nr = PKMAP_NR(vaddr); 276 switch (--pkmap_count[nr]) {
|
H A D | mincore.c | 93 unsigned long nr = (end - addr) >> PAGE_SHIFT; local 100 for (i = 0; i < nr; i++, pgoff++) 103 for (i = 0; i < nr; i++)
|
H A D | vmscan.c | 277 long nr; local 291 nr = atomic_long_xchg(&shrinker->nr_in_batch, 0); 293 total_scan = nr; 300 "delete nr=%ld\n", 309 * nr being built up so when a shrink that can do some work 310 * comes along it empties the entire cache due to nr >>> 321 * Avoid risking looping forever due to too large nr value: 328 trace_mm_shrink_slab_start(shrinker, shrink, nr, 359 trace_mm_shrink_slab_end(shrinker, shrink_ret, nr, new_nr); 1897 * nr[ 1899 get_scan_count(struct mem_cgroup_zone *mz, struct scan_control *sc, unsigned long *nr, int priority) argument 2082 unsigned long nr[NR_LRU_LISTS]; local 3122 int nr; local 3136 int nr; local [all...] |
H A D | swap.c | 227 release_pages(pvec->pages, pvec->nr, pvec->cold); 586 void release_pages(struct page **pages, int nr, int cold) argument 593 for (i = 0; i < nr; i++) { 748 pvec->nr = find_get_pages(mapping, start, nr_pages, pvec->pages); 756 pvec->nr = find_get_pages_tag(mapping, index, tag,
|
H A D | shmem.c | 391 pvec->nr = j; 412 pvec.nr = shmem_find_get_pages_and_swap(mapping, index, 414 if (!pvec.nr) 416 index = indices[pvec.nr - 1] + 1; 418 check_move_unevictable_pages(pvec.pages, pvec.nr); 445 pvec.nr = shmem_find_get_pages_and_swap(mapping, index, 448 if (!pvec.nr) 493 pvec.nr = shmem_find_get_pages_and_swap(mapping, index, 496 if (!pvec.nr) { 1230 unsigned long nr, re local [all...] |
H A D | readahead.c | 242 unsigned long max_sane_readahead(unsigned long nr) argument 244 return min(nr, (node_page_state(numa_node_id(), NR_INACTIVE_FILE)
|
H A D | filemap.c | 1113 unsigned long nr, ret; local 1150 * the correct value for "nr", which means the zero-filled 1162 /* nr is the maximum number of bytes to copy from this page */ 1163 nr = PAGE_CACHE_SIZE; 1165 nr = ((isize - 1) & ~PAGE_CACHE_MASK) + 1; 1166 if (nr <= offset) { 1171 nr = nr - offset; 1198 ret = actor(desc, page, offset, nr); 1205 if (ret == nr 1490 do_readahead(struct address_space *mapping, struct file *filp, pgoff_t index, unsigned long nr) argument [all...] |
H A D | memory-failure.c | 243 int nr; local 249 nr = shrink_slab(&shrink, 1000, 1000); 252 } while (nr > 10);
|
H A D | memory.c | 213 batch->nr = 0; 235 tlb->local.nr = 0; 260 free_pages_and_swap_cache(batch->pages, batch->nr); 261 batch->nr = 0; 304 batch->pages[batch->nr++] = page; 305 if (batch->nr == batch->max) { 310 VM_BUG_ON(batch->nr > batch->max); 312 return batch->max - batch->nr; 348 for (i = 0; i < batch->nr; i++) 385 (*batch)->nr [all...] |
H A D | slab.c | 964 int nr = min3(from->avail, max, to->limit - to->avail); local 966 if (!nr) 969 memcpy(to->entry + to->avail, from->entry + from->avail -nr, 970 sizeof(void *) *nr); 972 from->avail -= nr; 973 to->avail += nr; 974 return nr;
|
H A D | slub.c | 864 int nr = 0; local 870 while (fp && nr <= page->objects) { 890 nr++; 903 if (page->inuse != page->objects - nr) { 905 "counted were %d", page->inuse, page->objects - nr); 906 page->inuse = page->objects - nr;
|
H A D | hugetlb.c | 1875 unsigned int nr = 0; local 1878 nr += array[node]; 1880 return nr;
|
H A D | page_alloc.c | 2772 unsigned long nr[MAX_ORDER], flags, order, total = 0; local 2781 nr[order] = zone->free_area[order].nr_free; 2782 total += nr[order] << order; 2786 printk("%lu*%lukB ", nr[order], K(1UL) << order);
|