Searched refs:nr (Results 1 - 17 of 17) sorted by relevance

/mm/
H A Dquicklist.c63 void quicklist_trim(int nr, void (*dtor)(void *), argument
69 q = &get_cpu_var(quicklist)[nr];
78 void *p = quicklist_alloc(nr, 0, NULL);
H A Dfilemap_xip.c77 unsigned long nr, left; local
82 /* nr is the maximum number of bytes to copy from this page */
83 nr = PAGE_CACHE_SIZE;
87 nr = ((isize - 1) & ~PAGE_CACHE_MASK) + 1;
88 if (nr <= offset) {
92 nr = nr - offset;
93 if (nr > len - copied)
94 nr = len - copied;
123 left = __copy_to_user(buf+copied, xip_mem+offset, nr);
[all...]
H A Dswap_state.c238 void free_pages_and_swap_cache(struct page **pages, int nr) argument
243 while (nr) {
244 int todo = min(nr, PAGEVEC_SIZE);
251 nr -= todo;
H A Dvmalloc.c91 unsigned long end, pgprot_t prot, struct page **pages, int *nr)
96 * nr is a running index into the array which helps higher level
104 struct page *page = pages[*nr];
111 (*nr)++;
117 unsigned long end, pgprot_t prot, struct page **pages, int *nr)
127 if (vmap_pte_range(pmd, addr, next, prot, pages, nr))
134 unsigned long end, pgprot_t prot, struct page **pages, int *nr)
144 if (vmap_pmd_range(pud, addr, next, prot, pages, nr))
163 int nr = 0; local
169 err = vmap_pud_range(pgd, addr, next, prot, pages, &nr);
90 vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) argument
116 vmap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) argument
133 vmap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) argument
585 int nr = 0; local
2542 unsigned int nr, *counters = m->private; local
[all...]
H A Dhighmem.c262 unsigned long nr; local
269 nr = PKMAP_NR(vaddr);
276 switch (--pkmap_count[nr]) {
H A Dmincore.c93 unsigned long nr = (end - addr) >> PAGE_SHIFT; local
100 for (i = 0; i < nr; i++, pgoff++)
103 for (i = 0; i < nr; i++)
H A Dvmscan.c277 long nr; local
291 nr = atomic_long_xchg(&shrinker->nr_in_batch, 0);
293 total_scan = nr;
300 "delete nr=%ld\n",
309 * nr being built up so when a shrink that can do some work
310 * comes along it empties the entire cache due to nr >>>
321 * Avoid risking looping forever due to too large nr value:
328 trace_mm_shrink_slab_start(shrinker, shrink, nr,
359 trace_mm_shrink_slab_end(shrinker, shrink_ret, nr, new_nr);
1897 * nr[
1899 get_scan_count(struct mem_cgroup_zone *mz, struct scan_control *sc, unsigned long *nr, int priority) argument
2082 unsigned long nr[NR_LRU_LISTS]; local
3122 int nr; local
3136 int nr; local
[all...]
H A Dswap.c227 release_pages(pvec->pages, pvec->nr, pvec->cold);
586 void release_pages(struct page **pages, int nr, int cold) argument
593 for (i = 0; i < nr; i++) {
748 pvec->nr = find_get_pages(mapping, start, nr_pages, pvec->pages);
756 pvec->nr = find_get_pages_tag(mapping, index, tag,
H A Dshmem.c391 pvec->nr = j;
412 pvec.nr = shmem_find_get_pages_and_swap(mapping, index,
414 if (!pvec.nr)
416 index = indices[pvec.nr - 1] + 1;
418 check_move_unevictable_pages(pvec.pages, pvec.nr);
445 pvec.nr = shmem_find_get_pages_and_swap(mapping, index,
448 if (!pvec.nr)
493 pvec.nr = shmem_find_get_pages_and_swap(mapping, index,
496 if (!pvec.nr) {
1230 unsigned long nr, re local
[all...]
H A Dreadahead.c242 unsigned long max_sane_readahead(unsigned long nr) argument
244 return min(nr, (node_page_state(numa_node_id(), NR_INACTIVE_FILE)
H A Dfilemap.c1113 unsigned long nr, ret; local
1150 * the correct value for "nr", which means the zero-filled
1162 /* nr is the maximum number of bytes to copy from this page */
1163 nr = PAGE_CACHE_SIZE;
1165 nr = ((isize - 1) & ~PAGE_CACHE_MASK) + 1;
1166 if (nr <= offset) {
1171 nr = nr - offset;
1198 ret = actor(desc, page, offset, nr);
1205 if (ret == nr
1490 do_readahead(struct address_space *mapping, struct file *filp, pgoff_t index, unsigned long nr) argument
[all...]
H A Dmemory-failure.c243 int nr; local
249 nr = shrink_slab(&shrink, 1000, 1000);
252 } while (nr > 10);
H A Dmemory.c213 batch->nr = 0;
235 tlb->local.nr = 0;
260 free_pages_and_swap_cache(batch->pages, batch->nr);
261 batch->nr = 0;
304 batch->pages[batch->nr++] = page;
305 if (batch->nr == batch->max) {
310 VM_BUG_ON(batch->nr > batch->max);
312 return batch->max - batch->nr;
348 for (i = 0; i < batch->nr; i++)
385 (*batch)->nr
[all...]
H A Dslab.c964 int nr = min3(from->avail, max, to->limit - to->avail); local
966 if (!nr)
969 memcpy(to->entry + to->avail, from->entry + from->avail -nr,
970 sizeof(void *) *nr);
972 from->avail -= nr;
973 to->avail += nr;
974 return nr;
H A Dslub.c864 int nr = 0; local
870 while (fp && nr <= page->objects) {
890 nr++;
903 if (page->inuse != page->objects - nr) {
905 "counted were %d", page->inuse, page->objects - nr);
906 page->inuse = page->objects - nr;
H A Dhugetlb.c1875 unsigned int nr = 0; local
1878 nr += array[node];
1880 return nr;
H A Dpage_alloc.c2772 unsigned long nr[MAX_ORDER], flags, order, total = 0; local
2781 nr[order] = zone->free_area[order].nr_free;
2782 total += nr[order] << order;
2786 printk("%lu*%lukB ", nr[order], K(1UL) << order);

Completed in 508 milliseconds