/fs/squashfs/ |
H A D | page_actor.c | 32 if (actor->next_page == actor->pages) 44 int pages, int length) 51 actor->length = length ? : pages * PAGE_CACHE_SIZE; 53 actor->pages = pages; 73 return actor->pageaddr = actor->next_page == actor->pages ? NULL : 84 int pages, int length) 91 actor->length = length ? : pages * PAGE_CACHE_SIZE; 93 actor->pages = pages; 43 squashfs_page_actor_init(void **buffer, int pages, int length) argument 83 squashfs_page_actor_init_special(struct page **page, int pages, int length) argument [all...] |
H A D | page_actor.h | 14 int pages; member in struct:squashfs_page_actor 20 int pages, int length) 27 actor->length = length ? : pages * PAGE_CACHE_SIZE; 29 actor->pages = pages; 42 return actor->next_page == actor->pages ? NULL : 60 int pages; member in struct:squashfs_page_actor 19 squashfs_page_actor_init(void **page, int pages, int length) argument
|
H A D | file_direct.c | 24 int pages, struct page **page); 37 int i, n, pages, missing_pages, bytes, res = -ENOMEM; local 45 pages = end_index - start_index + 1; 47 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); 53 * page cache pages appropriately within the decompressor 55 actor = squashfs_page_actor_init_special(page, pages, 0); 59 /* Try to grab all the pages covered by the Squashfs block */ 60 for (missing_pages = 0, i = 0, n = start_index; i < pages; i++, n++) { 79 * Couldn't get one or more pages, this page has either 85 res = squashfs_read_cache(target_page, block, bsize, pages, 140 squashfs_read_cache(struct page *target_page, u64 block, int bsize, int pages, struct page **page) argument [all...] |
H A D | cache.c | 219 for (j = 0; j < cache->pages; j++) 258 cache->pages = block_size >> PAGE_CACHE_SHIFT; 259 cache->pages = cache->pages ? cache->pages : 1; 271 entry->data = kcalloc(cache->pages, sizeof(void *), GFP_KERNEL); 277 for (j = 0; j < cache->pages; j++) { 286 cache->pages, 0); 418 int pages = (length + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; local 427 data = kcalloc(pages, sizeo [all...] |
/fs/9p/ |
H A D | cache.h | 50 struct list_head *pages, 75 struct list_head *pages, 78 return __v9fs_readpages_from_fscache(inode, mapping, pages, 131 struct list_head *pages, 73 v9fs_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument 129 v9fs_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument
|
H A D | cache.c | 172 ClearPageFsCache(pvec.pages[loop]); 174 first = pvec.pages[nr_pages - 1]->index + 1; 316 * Returns 0 if the pages are in cache and a BIO is submitted, 317 * 1 if the pages are not in cache and -error otherwise. 349 * __v9fs_readpages_from_fscache - read multiple pages from cache 351 * Returns 0 if the pages are in cache and a BIO is submitted, 352 * 1 if the pages are not in cache and -error otherwise. 357 struct list_head *pages, 363 p9_debug(P9_DEBUG_FSC, "inode %p pages %u\n", inode, *nr_pages); 368 mapping, pages, nr_page 355 __v9fs_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument [all...] |
H A D | vfs_addr.c | 104 * v9fs_vfs_readpages - read a set of pages from 9P 108 * @pages: list of pages to read 109 * @nr_pages: count of pages to read 114 struct list_head *pages, unsigned nr_pages) 122 ret = v9fs_readpages_from_fscache(inode, mapping, pages, &nr_pages); 126 ret = read_cache_pages(mapping, pages, (void *)v9fs_vfs_readpage, filp); 113 v9fs_vfs_readpages(struct file *filp, struct address_space *mapping, struct list_head *pages, unsigned nr_pages) argument
|
/fs/proc/ |
H A D | meminfo.c | 32 unsigned long pages[NR_LRU_LISTS]; local 52 pages[lru] = global_page_state(NR_LRU_BASE + lru); 71 pagecache = pages[LRU_ACTIVE_FILE] + pages[LRU_INACTIVE_FILE]; 148 K(pages[LRU_ACTIVE_ANON] + pages[LRU_ACTIVE_FILE]), 149 K(pages[LRU_INACTIVE_ANON] + pages[LRU_INACTIVE_FILE]), 150 K(pages[LRU_ACTIVE_ANON]), 151 K(pages[LRU_INACTIVE_ANO [all...] |
/fs/afs/ |
H A D | file.c | 28 struct list_head *pages, unsigned nr_pages); 240 * read a set of pages 243 struct list_head *pages, unsigned nr_pages) 260 /* attempt to read as many of the pages as possible */ 264 pages, 274 /* all pages are being read from the cache */ 276 BUG_ON(!list_empty(pages)); 281 /* there were pages that couldn't be read from the cache */ 292 /* load the missing pages from the network */ 293 ret = read_cache_pages(mapping, pages, afs_page_fille 242 afs_readpages(struct file *file, struct address_space *mapping, struct list_head *pages, unsigned nr_pages) argument [all...] |
H A D | rxrpc.c | 271 * attach the data from a bunch of pages on an inode to a call 276 struct page *pages[8]; local 290 if (count > ARRAY_SIZE(pages)) 291 count = ARRAY_SIZE(pages); 292 n = find_get_pages_contig(call->mapping, first, count, pages); 303 iov->iov_base = kmap(pages[loop]) + offset; 319 kunmap(pages[loop]); 326 put_page(pages[loop]);
|
H A D | write.c | 276 * kill all the pages in the given range 296 first, count, pv.pages); 300 ClearPageUptodate(pv.pages[loop]); 302 SetPageError(pv.pages[loop]); 303 end_page_writeback(pv.pages[loop]); 314 * pages also covered by the same writeback record 319 struct page *pages[8], *page; local 333 /* find all consecutive lockable dirty pages, stopping when we find a 343 if (n > ARRAY_SIZE(pages)) 344 n = ARRAY_SIZE(pages); [all...] |
/fs/btrfs/tests/ |
H A D | extent-io-tests.c | 32 struct page *pages[16]; local 43 ARRAY_SIZE(pages)), pages); 46 !PageLocked(pages[i])) 48 if (flags & PROCESS_UNLOCK && PageLocked(pages[i])) 49 unlock_page(pages[i]); 50 page_cache_release(pages[i]); 52 page_cache_release(pages[i]); 88 * First go through and create and mark all of our pages dirty, we pin 89 * everything to make sure our pages do [all...] |
/fs/ceph/ |
H A D | cache.h | 45 struct list_head *pages, 85 struct list_head *pages) 88 return fscache_readpages_cancel(ci->fscache, pages); 121 struct page *pages) 133 struct list_head *pages, 172 struct list_head *pages) 84 ceph_fscache_readpages_cancel(struct inode *inode, struct list_head *pages) argument 120 ceph_fscache_uncache_page(struct inode *inode, struct page *pages) argument 131 ceph_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument 171 ceph_fscache_readpages_cancel(struct inode *inode, struct list_head *pages) argument
|
H A D | cache.c | 166 ClearPageFsCache(pvec.pages[loop]); 168 first = pvec.pages[nr_pages - 1]->index + 1; 281 struct list_head *pages, 290 ret = fscache_read_or_alloc_pages(ci->fscache, mapping, pages, nr_pages, 295 case 0: /* All pages found */ 298 case -ENOBUFS: /* Some pages were not found, and can't be */ 299 case -ENODATA: /* some pages were not found */ 279 ceph_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument
|
/fs/cifs/ |
H A D | fscache.h | 79 struct list_head *pages, 83 return __cifs_readpages_from_fscache(inode, mapping, pages, 96 struct list_head *pages) 99 return __cifs_fscache_readpages_cancel(inode, pages); 133 struct list_head *pages, 143 struct list_head *pages) 77 cifs_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument 95 cifs_fscache_readpages_cancel(struct inode *inode, struct list_head *pages) argument 131 cifs_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument 142 cifs_fscache_readpages_cancel(struct inode *inode, struct list_head *pages) argument
|
H A D | fscache.c | 182 * Retrieve a set of pages from FS-Cache 186 struct list_head *pages, 194 pages, nr_pages, 199 case 0: /* read submitted to the cache for all pages */ 203 case -ENOBUFS: /* some pages are not cached and can't be */ 204 case -ENODATA: /* some pages are not cached */ 226 void __cifs_fscache_readpages_cancel(struct inode *inode, struct list_head *pages) argument 230 fscache_readpages_cancel(CIFS_I(inode)->fscache, pages); 184 __cifs_readpages_from_fscache(struct inode *inode, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages) argument
|
/fs/nfs/ |
H A D | nfs3acl.c | 17 struct page *pages[NFSACL_MAXPAGES] = { }; local 20 /* The xdr layer may allocate pages here. */ 21 .pages = pages, 61 /* pages may have been allocated at the xdr layer. */ 62 for (count = 0; count < NFSACL_MAXPAGES && args.pages[count]; count++) 63 __free_page(args.pages[count]); 122 struct page *pages[NFSACL_MAXPAGES]; local 127 .pages = pages, [all...] |
/fs/ramfs/ |
H A D | file-nommu.c | 57 * add a contiguous set of pages into a ramfs inode when it's truncated from 64 struct page *pages; local 80 /* allocate enough contiguous pages to be able to satisfy the 82 pages = alloc_pages(mapping_gfp_mask(inode->i_mapping), order); 83 if (!pages) 86 /* split the high-order page into an array of single pages */ 90 split_page(pages, order); 92 /* trim off any pages we don't actually require */ 94 __free_page(pages + loop); 98 data = page_address(pages); 205 struct page **pages = NULL, **ptr, *page; local [all...] |
/fs/btrfs/ |
H A D | lzo.c | 90 struct page **pages, 134 pages[0] = out_page; 159 /* copy bytes from the working buffer into the pages */ 202 pages[nr_pages++] = out_page; 236 cpage_out = kmap(pages[0]); 239 kunmap(pages[0]); 315 /* copy bytes from the pages into the working buffer */ 87 lzo_compress_pages(struct list_head *ws, struct address_space *mapping, u64 start, unsigned long len, struct page **pages, unsigned long nr_dest_pages, unsigned long *out_pages, unsigned long *total_in, unsigned long *total_out, unsigned long max_out) argument
|
H A D | zlib.c | 77 struct page **pages, 115 pages[0] = out_page; 157 pages[nr_pages] = out_page; 74 zlib_compress_pages(struct list_head *ws, struct address_space *mapping, u64 start, unsigned long len, struct page **pages, unsigned long nr_dest_pages, unsigned long *out_pages, unsigned long *total_in, unsigned long *total_out, unsigned long max_out) argument
|
/fs/cachefiles/ |
H A D | rdwr.c | 19 * detect wake up events generated by the unlocking of pages in which we're 21 * - we use this to detect read completion of backing pages 146 * copy data from backing pages to netfs pages to complete a read operation 463 * read the corresponding pages to the given set from the backing file 464 * - any uncertain pages are simply discarded, to be tried again another time 675 * read a list of pages from the cache or allocate blocks in which to store 679 struct list_head *pages, 729 list_for_each_entry_safe(page, _n, pages, lru) { 765 if (list_empty(pages)) 678 cachefiles_read_or_alloc_pages(struct fscache_retrieval *op, struct list_head *pages, unsigned *nr_pages, gfp_t gfp) argument 833 cachefiles_allocate_pages(struct fscache_retrieval *op, struct list_head *pages, unsigned *nr_pages, gfp_t gfp) argument [all...] |
/fs/fscache/ |
H A D | page.c | 518 * -ENOMEM - out of memory, some pages may be being read 519 * -ERESTARTSYS - interrupted, some pages may be being read 521 * pages not being read 523 * the pages 524 * 0 - dispatched a read on all pages 529 * any pages for which a read is dispatched will be removed from pages and 534 struct list_head *pages, 559 ASSERT(!list_empty(pages)); 605 op, pages, nr_page 532 __fscache_read_or_alloc_pages(struct fscache_cookie *cookie, struct address_space *mapping, struct list_head *pages, unsigned *nr_pages, fscache_rw_complete_t end_io_func, void *context, gfp_t gfp) argument 743 __fscache_readpages_cancel(struct fscache_cookie *cookie, struct list_head *pages) argument [all...] |
/fs/hpfs/ |
H A D | file.c | 123 struct list_head *pages, unsigned nr_pages) 125 return mpage_readpages(mapping, pages, nr_pages, hpfs_get_block); 122 hpfs_readpages(struct file *file, struct address_space *mapping, struct list_head *pages, unsigned nr_pages) argument
|
/fs/isofs/ |
H A D | compress.c | 40 * to one zisofs block. Store the data in the @pages array with @pcount 45 struct page **pages, unsigned poffset, 71 if (!pages[i]) 73 memset(page_address(pages[i]), 0, PAGE_CACHE_SIZE); 74 flush_dcache_page(pages[i]); 75 SetPageUptodate(pages[i]); 121 if (pages[curpage]) { 122 stream.next_out = page_address(pages[curpage]) 174 if (pages[curpage]) { 175 flush_dcache_page(pages[curpag 43 zisofs_uncompress_block(struct inode *inode, loff_t block_start, loff_t block_end, int pcount, struct page **pages, unsigned poffset, int *errp) argument 199 zisofs_fill_pages(struct inode *inode, int full_page, int pcount, struct page **pages) argument 307 struct page *pages[max_t(unsigned, zisofs_pages_per_cblock, 1)]; local [all...] |
/fs/jfs/ |
H A D | inode.c | 266 * in 4K pages - see jfs_write.c 299 struct list_head *pages, unsigned nr_pages) 301 return mpage_readpages(mapping, pages, nr_pages, jfs_get_block); 298 jfs_readpages(struct file *file, struct address_space *mapping, struct list_head *pages, unsigned nr_pages) argument
|