/include/linux/ |
H A D | mm_inline.h | 7 * page_is_file_cache - should the page be on a file LRU or anon LRU? 8 * @page: the page to test 10 * Returns 1 if @page is page cache page backed by a regular filesystem, 11 * or 0 if @page is anonymous, tmpfs or otherwise ram or swap backed. 12 * Used by functions that manipulate the LRU lists, to sort a page 15 * We would like to get this info without a page flag, but the state 16 * needs to survive until the page i 19 page_is_file_cache(struct page *page) argument 25 add_page_to_lru_list(struct zone *zone, struct page *page, enum lru_list lru) argument 35 del_page_from_lru_list(struct zone *zone, struct page *page, enum lru_list lru) argument 50 page_lru_base_type(struct page *page) argument 64 page_off_lru(struct page *page) argument 88 page_lru(struct page *page) argument [all...] |
H A D | quicklist.h | 17 void *page; member in struct:quicklist 31 * the first cacheline of the page itself. There is minmal overhead involved. 39 p = q->page; 41 q->page = p[0]; 56 struct page *page) 61 *(void **)p = q->page; 62 q->page = p; 73 struct page *page) 55 __quicklist_free(int nr, void (*dtor)(void *), void *p, struct page *page) argument 72 quicklist_free_page(int nr, void (*dtor)(void *), struct page *page) argument [all...] |
H A D | pagevec.h | 14 struct page; 20 struct page *pages[PAGEVEC_SIZE]; 53 * Add a page to a pagevec. Returns the number of slots still available. 55 static inline unsigned pagevec_add(struct pagevec *pvec, struct page *page) argument 57 pvec->pages[pvec->nr++] = page;
|
H A D | migrate.h | 8 typedef struct page *new_page_t(struct page *, unsigned long private, int **); 14 struct page *, struct page *, enum migrate_mode); 23 struct page *, struct page *); 30 extern void migrate_page_copy(struct page *newpage, struct page *page); 32 struct page *newpag 53 migrate_page_copy(struct page *newpage, struct page *page) argument 56 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) argument [all...] |
H A D | pid_namespace.h | 13 void *page; member in struct:pidmap
|
H A D | async_tx.h | 77 * @scribble: caller provided space for dma/page address conversions 129 enum dma_transaction_type tx_type, struct page **dst, 130 int dst_count, struct page **src, int src_count, 151 struct page *page; member in union:__anon132 172 async_xor(struct page *dest, struct page **src_list, unsigned int offset, 176 async_xor_val(struct page *dest, struct page **src_list, unsigned int offset, 181 async_memcpy(struct page *des [all...] |
H A D | cleancache.h | 11 * cleancache requires every file with a page in cleancache to have a 28 pgoff_t, struct page *); 30 pgoff_t, struct page *); 40 extern int __cleancache_get_page(struct page *); 41 extern void __cleancache_put_page(struct page *); 42 extern void __cleancache_invalidate_page(struct address_space *, struct page *); 48 static inline bool cleancache_fs_enabled(struct page *page) argument 50 return page->mapping->host->i_sb->cleancache_poolid >= 0; 87 static inline int cleancache_get_page(struct page *pag argument 96 cleancache_put_page(struct page *page) argument 102 cleancache_invalidate_page(struct address_space *mapping, struct page *page) argument [all...] |
H A D | dm-io.h | 25 struct page *page; member in struct:page_list
|
H A D | dma-debug.h | 37 extern void debug_dma_map_page(struct device *dev, struct page *page, 101 static inline void debug_dma_map_page(struct device *dev, struct page *page, argument
|
H A D | huge_mm.h | 15 extern struct page *follow_trans_huge_pmd(struct mm_struct *mm, 49 extern pmd_t *page_check_address_pmd(struct page *page, 92 extern int split_huge_page(struct page *page); 138 static inline int hpage_nr_pages(struct page *page) argument 140 if (unlikely(PageTransHuge(page))) 144 static inline struct page *compound_trans_head(struct page *pag argument 171 split_huge_page(struct page *page) argument [all...] |
H A D | kmemcheck.h | 11 void kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node); 12 void kmemcheck_free_shadow(struct page *page, int order); 17 void kmemcheck_pagealloc_alloc(struct page *p, unsigned int order, 20 void kmemcheck_show_pages(struct page *p, unsigned int n); 21 void kmemcheck_hide_pages(struct page *p, unsigned int n); 23 bool kmemcheck_page_is_tracked(struct page *p); 30 void kmemcheck_mark_unallocated_pages(struct page *p, unsigned int n); 31 void kmemcheck_mark_uninitialized_pages(struct page * 93 kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node) argument 98 kmemcheck_free_shadow(struct page *page, int order) argument [all...] |
H A D | ksm.h | 19 struct page *ksm_does_need_to_copy(struct page *page, 42 * A KSM page is one of those write-protected "shared pages" or "merged pages" 43 * which KSM maps into multiple mms, wherever identical anonymous page content 44 * is found in VM_MERGEABLE vmas. It's a PageAnon page, pointing not to any 45 * anon_vma, but to that page's node of the stable tree. 47 static inline int PageKsm(struct page *page) argument 49 return ((unsigned long)page 53 page_stable_node(struct page *page) argument 58 set_page_stable_node(struct page *page, struct stable_node *stable_node) argument 76 ksm_might_need_to_copy(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 104 PageKsm(struct page *page) argument 116 ksm_might_need_to_copy(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 122 page_referenced_ksm(struct page *page, struct mem_cgroup *memcg, unsigned long *vm_flags) argument 128 try_to_unmap_ksm(struct page *page, enum ttu_flags flags) argument 133 rmap_walk_ksm(struct page *page, int (*rmap_one)(struct page*, struct vm_area_struct *, unsigned long, void *), void *arg) argument [all...] |
H A D | memory_hotplug.h | 9 struct page; 17 * Types for free bootmem stored in page->lru.next. These have to be in 71 extern int add_one_highpage(struct page *page, int pfn, int bad_ppro); 76 typedef void (*online_page_callback_t)(struct page *page); 81 extern void __online_page_set_limits(struct page *page); 82 extern void __online_page_increment_counters(struct page *page); 168 put_page_bootmem(struct page *page) argument [all...] |
H A D | page_cgroup.h | 8 PCG_MIGRATION, /* under page migration */ 20 * A page_cgroup page is associated with every page descriptor. The 22 * All page cgroups are allocated at boot or memory hotplug event, 23 * then the page cgroup for pfn always exists. 44 struct page_cgroup *lookup_page_cgroup(struct page *page); 45 struct page *lookup_cgroup_page(struct page_cgroup *pc); 92 static inline struct page_cgroup *lookup_page_cgroup(struct page *page) argument [all...] |
H A D | pipe_fs_i.h | 6 #define PIPE_BUF_FLAG_LRU 0x01 /* page is on the LRU */ 8 #define PIPE_BUF_FLAG_GIFT 0x04 /* page is a gift */ 13 * @page: the page containing the data for the pipe buffer 14 * @offset: offset of data inside the @page 15 * @len: length of data inside the @page 21 struct page *page; member in struct:pipe_buffer 34 * @tmp_page: cached released page 53 struct page *tmp_pag [all...] |
H A D | swapops.h | 104 static inline swp_entry_t make_migration_entry(struct page *page, int write) argument 106 BUG_ON(!PageLocked(page)); 108 page_to_pfn(page)); 122 static inline struct page *migration_entry_to_page(swp_entry_t entry) 124 struct page *p = pfn_to_page(swp_offset(entry)); 127 * corresponding page is locked 142 #define make_migration_entry(page, write) swp_entry(0, 0) 162 static inline swp_entry_t make_hwpoison_entry(struct page *page) argument 174 make_hwpoison_entry(struct page *page) argument [all...] |
H A D | vmstat.h | 88 * Zone based page accounting with per cpu differentials. 191 void __inc_zone_page_state(struct page *, enum zone_stat_item); 192 void __dec_zone_page_state(struct page *, enum zone_stat_item); 195 void inc_zone_page_state(struct page *, enum zone_stat_item); 196 void dec_zone_page_state(struct page *, enum zone_stat_item); 228 static inline void __inc_zone_page_state(struct page *page, argument 231 __inc_zone_state(page_zone(page), item); 240 static inline void __dec_zone_page_state(struct page *page, argument [all...] |
H A D | binfmts.h | 35 struct page *page[MAX_ARG_PAGES]; member in struct:linux_binprm
|
H A D | highmem.h | 14 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument 20 static inline void flush_kernel_dcache_page(struct page *page) argument 49 static inline void *kmap(struct page *page) argument 52 return page_address(page); 55 static inline void kunmap(struct page *page) argument 59 static inline void *kmap_atomic(struct page *pag argument 242 zero_user_segment(struct page *page, unsigned start, unsigned end) argument 248 zero_user(struct page *page, unsigned start, unsigned size) argument 254 memclear_highpage_flush(struct page *page, unsigned int offset, unsigned int size) argument [all...] |
H A D | rmap.h | 15 * an anonymous page pointing to this anon_vma needs to be unmapped: 19 * in mprotect), the mapping field of an anonymous page cannot point 24 * the anon_vma object itself: we're guaranteed no page can be 32 * guarantee that the vma of page tables will exist for 84 static inline struct anon_vma *page_anon_vma(struct page *page) argument 86 if (((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) != 89 return page_rmapping(page); 133 struct anon_vma *page_get_anon_vma(struct page *page); 151 page_dup_rmap(struct page *page) argument 188 page_check_address(struct page *page, struct mm_struct *mm, unsigned long address, spinlock_t **ptlp, int sync) argument 237 page_referenced(struct page *page, int is_locked, struct mem_cgroup *memcg, unsigned long *vm_flags) argument 247 page_mkclean(struct page *page) argument [all...] |
H A D | scatterlist.h | 22 * scatterlist struct. We use that to place the page pointer AND encode 38 * We overload the LSB of the page pointer to indicate whether it's 48 * sg_assign_page - Assign a given page to an SG entry 50 * @page: The page 53 * Assign page to sg entry. Also see sg_set_page(), the most commonly used 57 static inline void sg_assign_page(struct scatterlist *sg, struct page *page) argument 65 BUG_ON((unsigned long) page & 0x03); 70 sg->page_link = page_link | (unsigned long) page; 87 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len, unsigned int offset) argument 252 struct page *page; /* currently mapped page */ member in struct:sg_mapping_iter [all...] |
/include/asm-generic/ |
H A D | dma-mapping-common.h | 71 static inline dma_addr_t dma_map_page(struct device *dev, struct page *page, argument 78 kmemcheck_mark_initialized(page_address(page) + offset, size); 80 addr = ops->map_page(dev, page, offset, size, dir, NULL); 81 debug_dma_map_page(dev, page, offset, size, dir, addr, false);
|
H A D | pci-dma-compat.h | 43 pci_map_page(struct pci_dev *hwdev, struct page *page, argument 46 return dma_map_page(hwdev == NULL ? NULL : &hwdev->dev, page, offset, size, (enum dma_data_direction)direction);
|
H A D | tlb.h | 24 * Semi RCU freeing of the page directories. 28 * gup_fast() and other software pagetable walkers do a lockless page-table 29 * walk and therefore needs some synchronization with the freeing of the page 34 * since we unlink the page, flush TLBs, free the page. Since the disabling of 36 * freed page. 66 * If we can't allocate a page to make a big batch of page pointers 75 struct page *pages[0]; 96 struct page *__page 123 tlb_remove_page(struct mmu_gather *tlb, struct page *page) argument [all...] |
/include/drm/ttm/ |
H A D | ttm_bo_api.h | 51 * @fpfn: first valid page frame number to put the object 52 * @lpfn: last valid page frame number to put the object 280 * @page: The page when kmap'ing a single page. 292 struct page *page; member in struct:ttm_bo_kmap_obj 651 * @start_page: The first page to map.
|