Searched refs:newpage (Results 1 - 6 of 6) sorted by relevance
/mm/ |
H A D | migrate.c | 341 struct page *newpage, struct page *page, 389 get_page(newpage); /* add cache reference */ 391 SetPageSwapCache(newpage); 392 set_page_private(newpage, page_private(page)); 395 radix_tree_replace_slot(pslot, newpage); 415 __inc_zone_page_state(newpage, NR_FILE_PAGES); 418 __inc_zone_page_state(newpage, NR_SHMEM); 430 struct page *newpage, struct page *page) 458 get_page(newpage); 460 radix_tree_replace_slot(pslot, newpage); 340 migrate_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page, struct buffer_head *head, enum migrate_mode mode, int extra_count) argument 429 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) argument 519 migrate_page_copy(struct page *newpage, struct page *page) argument 594 migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) argument 618 buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) argument 716 fallback_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) argument 748 move_to_new_page(struct page *newpage, struct page *page, int remap_swapcache, enum migrate_mode mode) argument 797 __unmap_and_move(struct page *page, struct page *newpage, int force, enum migrate_mode mode) argument 942 struct page *newpage = get_new_page(page, private, &result); local 1586 struct page *newpage; local [all...] |
H A D | balloon_compaction.c | 196 int balloon_page_migrate(struct page *newpage, argument 203 * Block others from accessing the 'newpage' when we get around to 205 * holding a reference to the 'newpage' at this point. 207 BUG_ON(!trylock_page(newpage)); 211 unlock_page(newpage); 216 rc = balloon->migratepage(balloon, newpage, page, mode); 218 unlock_page(newpage);
|
H A D | internal.h | 254 static inline void mlock_migrate_page(struct page *newpage, struct page *page) argument 262 SetPageMlocked(newpage); 263 __mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages);
|
H A D | shmem.c | 967 struct page *oldpage, *newpage; local 981 newpage = shmem_alloc_page(gfp, info, index); 982 if (!newpage) 985 page_cache_get(newpage); 986 copy_highpage(newpage, oldpage); 987 flush_dcache_page(newpage); 989 __set_page_locked(newpage); 990 SetPageUptodate(newpage); 991 SetPageSwapBacked(newpage); 992 set_page_private(newpage, swap_inde [all...] |
H A D | ksm.c | 1956 void ksm_migrate_page(struct page *newpage, struct page *oldpage) argument 1961 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage); 1962 VM_BUG_ON_PAGE(newpage->mapping != oldpage->mapping, newpage); 1964 stable_node = page_stable_node(newpage); 1967 stable_node->kpfn = page_to_pfn(newpage); 1969 * newpage->mapping was set in advance; now we need smp_wmb()
|
H A D | memcontrol.c | 6603 * @newpage: page to transfer the charge to 6606 * Migrate the charge from @oldpage to @newpage. 6608 * Both pages must be locked, @newpage->mapping must be set up. 6610 void mem_cgroup_migrate(struct page *oldpage, struct page *newpage, argument 6617 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage); 6619 VM_BUG_ON_PAGE(!lrucare && PageLRU(newpage), newpage); 6620 VM_BUG_ON_PAGE(PageAnon(oldpage) != PageAnon(newpage), newpage); local 6622 newpage); local [all...] |
Completed in 94 milliseconds