/mm/ |
H A D | filemap_xip.c | 60 pgoff_t index, end_index; local 68 index = pos >> PAGE_CACHE_SHIFT; 84 if (index >= end_index) { 85 if (index > end_index) 96 error = mapping->a_ops->get_xip_mem(mapping, index, 0, 134 index += offset >> PAGE_CACHE_SHIFT; 334 unsigned long index; local 341 index = pos >> PAGE_CACHE_SHIFT; 346 status = a_ops->get_xip_mem(mapping, index, 0, 351 status = a_ops->get_xip_mem(mapping, index, 449 pgoff_t index = from >> PAGE_CACHE_SHIFT; local [all...] |
H A D | truncate.c | 29 * @offset: the index of the truncation point 145 (loff_t)page->index << PAGE_CACHE_SHIFT, 212 pgoff_t index; local 224 index = start; 225 while (index <= end && pagevec_lookup(&pvec, mapping, index, 226 min(end - index, (pgoff_t)PAGEVEC_SIZE - 1) + 1)) { 231 /* We rely upon deletion not changing page->index */ 232 index = page->index; 334 pgoff_t index = start; local 441 pgoff_t index; local [all...] |
H A D | swap.c | 733 * @start: The starting page index 754 pgoff_t *index, int tag, unsigned nr_pages) 756 pvec->nr = find_get_pages_tag(mapping, index, tag, 753 pagevec_lookup_tag(struct pagevec *pvec, struct address_space *mapping, pgoff_t *index, int tag, unsigned nr_pages) argument
|
H A D | vmstat.c | 571 * A fragmentation index only makes sense if an allocation of a requested 572 * size would fail. If that is true, the fragmentation index indicates 584 /* Fragmentation index only makes sense when a request would fail */ 597 /* Same as __fragmentation index but allocs contig_page_info on stack */ 1226 * Return an index indicating how much of the available free memory is 1251 int index; local 1259 index = unusable_free_index(order, &info); 1260 seq_printf(m, "%d.%03d ", index / 1000, index % 1000); 1267 * Display unusable free space index 1311 int index; local [all...] |
H A D | filemap.c | 128 radix_tree_delete(&mapping->page_tree, page->index); 130 /* Leave page->index set: truncation lookup relies upon it */ 266 pgoff_t index = start_byte >> PAGE_CACHE_SHIFT; local 276 while ((index <= end) && 277 (nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, 279 min(end - index, (pgoff_t)PAGEVEC_SIZE-1) + 1)) != 0) { 286 if (page->index > end) 406 pgoff_t offset = old->index; 411 new->index = offset; 438 * @offset: page index 765 find_or_create_page(struct address_space *mapping, pgoff_t index, gfp_t gfp_mask) argument 885 find_get_pages_contig(struct address_space *mapping, pgoff_t index, unsigned int nr_pages, struct page **pages) argument 960 find_get_pages_tag(struct address_space *mapping, pgoff_t *index, int tag, unsigned int nr_pages, struct page **pages) argument 1037 grab_cache_page_nowait(struct address_space *mapping, pgoff_t index) argument 1096 pgoff_t index; local 1490 do_readahead(struct address_space *mapping, struct file *filp, pgoff_t index, unsigned long nr) argument 1803 __read_cache_page(struct address_space *mapping, pgoff_t index, int (*filler)(void *, struct page *), void *data, gfp_t gfp) argument 1834 do_read_cache_page(struct address_space *mapping, pgoff_t index, int (*filler)(void *, struct page *), void *data, gfp_t gfp) argument 1886 read_cache_page_async(struct address_space *mapping, pgoff_t index, int (*filler)(void *, struct page *), void *data) argument 1918 read_cache_page_gfp(struct address_space *mapping, pgoff_t index, gfp_t gfp) argument 1940 read_cache_page(struct address_space *mapping, pgoff_t index, int (*filler)(void *, struct page *), void *data) argument 2336 grab_cache_page_write_begin(struct address_space *mapping, pgoff_t index, unsigned flags) argument [all...] |
H A D | memory.c | 684 pgoff_t index; local 710 index = linear_page_index(vma, addr); 719 "addr:%p vm_flags:%08lx anon_vma:%p mapping:%p index:%lx\n", 720 (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index); 1153 * Each page->index must be checked when 1157 (page->index < details->first_index || 1158 page->index > details->last_index)) 1168 addr) != page->index) 1170 pgoff_to_pte(page->index)); 2552 vmf.pgoff = old_page->index; [all...] |
H A D | page-writeback.c | 1617 * @start: starting page index 1618 * @end: ending page index (inclusive) 1681 pgoff_t index; local 1691 index = writeback_index; 1692 if (index == 0) 1698 index = wbc->range_start >> PAGE_CACHE_SHIFT; 1710 tag_pages_for_writeback(mapping, index, end); 1711 done_index = index; 1712 while (!done && (index <= end)) { 1715 nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, ta [all...] |
H A D | shmem.c | 106 static int shmem_getpage_gfp(struct inode *inode, pgoff_t index, 109 static inline int shmem_getpage(struct inode *inode, pgoff_t index, argument 112 return shmem_getpage_gfp(inode, index, pagep, sgp, 230 pgoff_t index, void *expected, void *replacement) 236 pslot = radix_tree_lookup_slot(&mapping->page_tree, index); 245 radix_tree_delete(&mapping->page_tree, index); 254 pgoff_t index, gfp_t gfp, void *expected) 266 page->index = index; 271 index, pag 229 shmem_radix_tree_replace(struct address_space *mapping, pgoff_t index, void *expected, void *replacement) argument 252 shmem_add_to_page_cache(struct page *page, struct address_space *mapping, pgoff_t index, gfp_t gfp, void *expected) argument 366 shmem_free_swap(struct address_space *mapping, pgoff_t index, void *radswap) argument 401 pgoff_t index = 0; local 437 pgoff_t index; local 611 pgoff_t index; local 707 pgoff_t index; local 798 shmem_swapin(swp_entry_t swap, gfp_t gfp, struct shmem_inode_info *info, pgoff_t index) argument 815 shmem_alloc_page(gfp_t gfp, struct shmem_inode_info *info, pgoff_t index) argument 838 shmem_swapin(swp_entry_t swap, gfp_t gfp, struct shmem_inode_info *info, pgoff_t index) argument 844 shmem_alloc_page(gfp_t gfp, struct shmem_inode_info *info, pgoff_t index) argument 865 shmem_getpage_gfp(struct inode *inode, pgoff_t index, struct page **pagep, enum sgp_type sgp, gfp_t gfp, int *fault_type) argument 1079 pgoff_t index; local 1187 pgoff_t index = pos >> PAGE_CACHE_SHIFT; local 1212 pgoff_t index; local 1356 pgoff_t index, end_index; local 2598 shmem_read_mapping_page_gfp(struct address_space *mapping, pgoff_t index, gfp_t gfp) argument [all...] |
H A D | slab.c | 307 * it. Mostly the same as what is in linux/slab.h except it returns an index. 1481 static void __init set_up_list3s(struct kmem_cache *cachep, int index) argument 1486 cachep->nodelists[node] = &initkmem_list3[index + node];
|
H A D | slub.c | 294 /* Determine object index from a given position */ 3254 * Conversion table for small slabs sizes / 8 to the index in the 3293 int index; local 3299 index = size_index[size_index_elem(size)]; 3301 index = fls(size - 1); 3305 return kmalloc_dma_caches[index]; 3308 return kmalloc_caches[index]; 3736 * handle the index determination for the smaller caches.
|