Searched defs:page (Results 1 - 25 of 50) sorted by relevance

12

/include/linux/
H A Dmm_inline.h7 * page_is_file_cache - should the page be on a file LRU or anon LRU?
8 * @page: the page to test
10 * Returns 1 if @page is page cache page backed by a regular filesystem,
11 * or 0 if @page is anonymous, tmpfs or otherwise ram or swap backed.
12 * Used by functions that manipulate the LRU lists, to sort a page
15 * We would like to get this info without a page flag, but the state
16 * needs to survive until the page i
19 page_is_file_cache(struct page *page) argument
25 add_page_to_lru_list(struct zone *zone, struct page *page, enum lru_list lru) argument
35 del_page_from_lru_list(struct zone *zone, struct page *page, enum lru_list lru) argument
50 page_lru_base_type(struct page *page) argument
64 page_off_lru(struct page *page) argument
88 page_lru(struct page *page) argument
[all...]
H A Dquicklist.h17 void *page; member in struct:quicklist
31 * the first cacheline of the page itself. There is minmal overhead involved.
39 p = q->page;
41 q->page = p[0];
56 struct page *page)
61 *(void **)p = q->page;
62 q->page = p;
73 struct page *page)
55 __quicklist_free(int nr, void (*dtor)(void *), void *p, struct page *page) argument
72 quicklist_free_page(int nr, void (*dtor)(void *), struct page *page) argument
[all...]
H A Dpagevec.h14 struct page;
20 struct page *pages[PAGEVEC_SIZE];
53 * Add a page to a pagevec. Returns the number of slots still available.
55 static inline unsigned pagevec_add(struct pagevec *pvec, struct page *page) argument
57 pvec->pages[pvec->nr++] = page;
H A Dmigrate.h8 typedef struct page *new_page_t(struct page *, unsigned long private, int **);
14 struct page *, struct page *, enum migrate_mode);
23 struct page *, struct page *);
30 extern void migrate_page_copy(struct page *newpage, struct page *page);
32 struct page *newpag
53 migrate_page_copy(struct page *newpage, struct page *page) argument
56 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) argument
[all...]
H A Dpid_namespace.h13 void *page; member in struct:pidmap
H A Dasync_tx.h77 * @scribble: caller provided space for dma/page address conversions
129 enum dma_transaction_type tx_type, struct page **dst,
130 int dst_count, struct page **src, int src_count,
151 struct page *page; member in union:__anon132
172 async_xor(struct page *dest, struct page **src_list, unsigned int offset,
176 async_xor_val(struct page *dest, struct page **src_list, unsigned int offset,
181 async_memcpy(struct page *des
[all...]
H A Dcleancache.h11 * cleancache requires every file with a page in cleancache to have a
28 pgoff_t, struct page *);
30 pgoff_t, struct page *);
40 extern int __cleancache_get_page(struct page *);
41 extern void __cleancache_put_page(struct page *);
42 extern void __cleancache_invalidate_page(struct address_space *, struct page *);
48 static inline bool cleancache_fs_enabled(struct page *page) argument
50 return page->mapping->host->i_sb->cleancache_poolid >= 0;
87 static inline int cleancache_get_page(struct page *pag argument
96 cleancache_put_page(struct page *page) argument
102 cleancache_invalidate_page(struct address_space *mapping, struct page *page) argument
[all...]
H A Ddm-io.h25 struct page *page; member in struct:page_list
H A Ddma-debug.h37 extern void debug_dma_map_page(struct device *dev, struct page *page,
101 static inline void debug_dma_map_page(struct device *dev, struct page *page, argument
H A Dhuge_mm.h15 extern struct page *follow_trans_huge_pmd(struct mm_struct *mm,
49 extern pmd_t *page_check_address_pmd(struct page *page,
92 extern int split_huge_page(struct page *page);
138 static inline int hpage_nr_pages(struct page *page) argument
140 if (unlikely(PageTransHuge(page)))
144 static inline struct page *compound_trans_head(struct page *pag argument
171 split_huge_page(struct page *page) argument
[all...]
H A Dkmemcheck.h11 void kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node);
12 void kmemcheck_free_shadow(struct page *page, int order);
17 void kmemcheck_pagealloc_alloc(struct page *p, unsigned int order,
20 void kmemcheck_show_pages(struct page *p, unsigned int n);
21 void kmemcheck_hide_pages(struct page *p, unsigned int n);
23 bool kmemcheck_page_is_tracked(struct page *p);
30 void kmemcheck_mark_unallocated_pages(struct page *p, unsigned int n);
31 void kmemcheck_mark_uninitialized_pages(struct page *
93 kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node) argument
98 kmemcheck_free_shadow(struct page *page, int order) argument
[all...]
H A Dksm.h19 struct page *ksm_does_need_to_copy(struct page *page,
42 * A KSM page is one of those write-protected "shared pages" or "merged pages"
43 * which KSM maps into multiple mms, wherever identical anonymous page content
44 * is found in VM_MERGEABLE vmas. It's a PageAnon page, pointing not to any
45 * anon_vma, but to that page's node of the stable tree.
47 static inline int PageKsm(struct page *page) argument
49 return ((unsigned long)page
53 page_stable_node(struct page *page) argument
58 set_page_stable_node(struct page *page, struct stable_node *stable_node) argument
76 ksm_might_need_to_copy(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
104 PageKsm(struct page *page) argument
116 ksm_might_need_to_copy(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
122 page_referenced_ksm(struct page *page, struct mem_cgroup *memcg, unsigned long *vm_flags) argument
128 try_to_unmap_ksm(struct page *page, enum ttu_flags flags) argument
133 rmap_walk_ksm(struct page *page, int (*rmap_one)(struct page*, struct vm_area_struct *, unsigned long, void *), void *arg) argument
[all...]
H A Dmemory_hotplug.h9 struct page;
17 * Types for free bootmem stored in page->lru.next. These have to be in
71 extern int add_one_highpage(struct page *page, int pfn, int bad_ppro);
76 typedef void (*online_page_callback_t)(struct page *page);
81 extern void __online_page_set_limits(struct page *page);
82 extern void __online_page_increment_counters(struct page *page);
168 put_page_bootmem(struct page *page) argument
[all...]
H A Dpage_cgroup.h8 PCG_MIGRATION, /* under page migration */
20 * A page_cgroup page is associated with every page descriptor. The
22 * All page cgroups are allocated at boot or memory hotplug event,
23 * then the page cgroup for pfn always exists.
44 struct page_cgroup *lookup_page_cgroup(struct page *page);
45 struct page *lookup_cgroup_page(struct page_cgroup *pc);
92 static inline struct page_cgroup *lookup_page_cgroup(struct page *page) argument
[all...]
H A Dpipe_fs_i.h6 #define PIPE_BUF_FLAG_LRU 0x01 /* page is on the LRU */
8 #define PIPE_BUF_FLAG_GIFT 0x04 /* page is a gift */
13 * @page: the page containing the data for the pipe buffer
14 * @offset: offset of data inside the @page
15 * @len: length of data inside the @page
21 struct page *page; member in struct:pipe_buffer
34 * @tmp_page: cached released page
53 struct page *tmp_pag
[all...]
H A Dswapops.h104 static inline swp_entry_t make_migration_entry(struct page *page, int write) argument
106 BUG_ON(!PageLocked(page));
108 page_to_pfn(page));
122 static inline struct page *migration_entry_to_page(swp_entry_t entry)
124 struct page *p = pfn_to_page(swp_offset(entry));
127 * corresponding page is locked
142 #define make_migration_entry(page, write) swp_entry(0, 0)
162 static inline swp_entry_t make_hwpoison_entry(struct page *page) argument
174 make_hwpoison_entry(struct page *page) argument
[all...]
H A Dvmstat.h88 * Zone based page accounting with per cpu differentials.
191 void __inc_zone_page_state(struct page *, enum zone_stat_item);
192 void __dec_zone_page_state(struct page *, enum zone_stat_item);
195 void inc_zone_page_state(struct page *, enum zone_stat_item);
196 void dec_zone_page_state(struct page *, enum zone_stat_item);
228 static inline void __inc_zone_page_state(struct page *page, argument
231 __inc_zone_state(page_zone(page), item);
240 static inline void __dec_zone_page_state(struct page *page, argument
[all...]
H A Dbinfmts.h35 struct page *page[MAX_ARG_PAGES]; member in struct:linux_binprm
H A Dhighmem.h14 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) argument
20 static inline void flush_kernel_dcache_page(struct page *page) argument
49 static inline void *kmap(struct page *page) argument
52 return page_address(page);
55 static inline void kunmap(struct page *page) argument
59 static inline void *kmap_atomic(struct page *pag argument
242 zero_user_segment(struct page *page, unsigned start, unsigned end) argument
248 zero_user(struct page *page, unsigned start, unsigned size) argument
254 memclear_highpage_flush(struct page *page, unsigned int offset, unsigned int size) argument
[all...]
H A Drmap.h15 * an anonymous page pointing to this anon_vma needs to be unmapped:
19 * in mprotect), the mapping field of an anonymous page cannot point
24 * the anon_vma object itself: we're guaranteed no page can be
32 * guarantee that the vma of page tables will exist for
84 static inline struct anon_vma *page_anon_vma(struct page *page) argument
86 if (((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) !=
89 return page_rmapping(page);
133 struct anon_vma *page_get_anon_vma(struct page *page);
151 page_dup_rmap(struct page *page) argument
188 page_check_address(struct page *page, struct mm_struct *mm, unsigned long address, spinlock_t **ptlp, int sync) argument
237 page_referenced(struct page *page, int is_locked, struct mem_cgroup *memcg, unsigned long *vm_flags) argument
247 page_mkclean(struct page *page) argument
[all...]
H A Dscatterlist.h22 * scatterlist struct. We use that to place the page pointer AND encode
38 * We overload the LSB of the page pointer to indicate whether it's
48 * sg_assign_page - Assign a given page to an SG entry
50 * @page: The page
53 * Assign page to sg entry. Also see sg_set_page(), the most commonly used
57 static inline void sg_assign_page(struct scatterlist *sg, struct page *page) argument
65 BUG_ON((unsigned long) page & 0x03);
70 sg->page_link = page_link | (unsigned long) page;
87 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len, unsigned int offset) argument
252 struct page *page; /* currently mapped page */ member in struct:sg_mapping_iter
[all...]
/include/asm-generic/
H A Ddma-mapping-common.h71 static inline dma_addr_t dma_map_page(struct device *dev, struct page *page, argument
78 kmemcheck_mark_initialized(page_address(page) + offset, size);
80 addr = ops->map_page(dev, page, offset, size, dir, NULL);
81 debug_dma_map_page(dev, page, offset, size, dir, addr, false);
H A Dpci-dma-compat.h43 pci_map_page(struct pci_dev *hwdev, struct page *page, argument
46 return dma_map_page(hwdev == NULL ? NULL : &hwdev->dev, page, offset, size, (enum dma_data_direction)direction);
H A Dtlb.h24 * Semi RCU freeing of the page directories.
28 * gup_fast() and other software pagetable walkers do a lockless page-table
29 * walk and therefore needs some synchronization with the freeing of the page
34 * since we unlink the page, flush TLBs, free the page. Since the disabling of
36 * freed page.
66 * If we can't allocate a page to make a big batch of page pointers
75 struct page *pages[0];
96 struct page *__page
123 tlb_remove_page(struct mmu_gather *tlb, struct page *page) argument
[all...]
/include/drm/ttm/
H A Dttm_bo_api.h51 * @fpfn: first valid page frame number to put the object
52 * @lpfn: last valid page frame number to put the object
280 * @page: The page when kmap'ing a single page.
292 struct page *page; member in struct:ttm_bo_kmap_obj
651 * @start_page: The first page to map.

Completed in 248 milliseconds

12