Searched refs:newpage (Results 1 - 6 of 6) sorted by relevance

/mm/
H A Dmigrate.c341 struct page *newpage, struct page *page,
389 get_page(newpage); /* add cache reference */
391 SetPageSwapCache(newpage);
392 set_page_private(newpage, page_private(page));
395 radix_tree_replace_slot(pslot, newpage);
415 __inc_zone_page_state(newpage, NR_FILE_PAGES);
418 __inc_zone_page_state(newpage, NR_SHMEM);
430 struct page *newpage, struct page *page)
458 get_page(newpage);
460 radix_tree_replace_slot(pslot, newpage);
340 migrate_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page, struct buffer_head *head, enum migrate_mode mode, int extra_count) argument
429 migrate_huge_page_move_mapping(struct address_space *mapping, struct page *newpage, struct page *page) argument
519 migrate_page_copy(struct page *newpage, struct page *page) argument
594 migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) argument
618 buffer_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) argument
716 fallback_migrate_page(struct address_space *mapping, struct page *newpage, struct page *page, enum migrate_mode mode) argument
748 move_to_new_page(struct page *newpage, struct page *page, int remap_swapcache, enum migrate_mode mode) argument
797 __unmap_and_move(struct page *page, struct page *newpage, int force, enum migrate_mode mode) argument
942 struct page *newpage = get_new_page(page, private, &result); local
1586 struct page *newpage; local
[all...]
H A Dballoon_compaction.c196 int balloon_page_migrate(struct page *newpage, argument
203 * Block others from accessing the 'newpage' when we get around to
205 * holding a reference to the 'newpage' at this point.
207 BUG_ON(!trylock_page(newpage));
211 unlock_page(newpage);
216 rc = balloon->migratepage(balloon, newpage, page, mode);
218 unlock_page(newpage);
H A Dinternal.h254 static inline void mlock_migrate_page(struct page *newpage, struct page *page) argument
262 SetPageMlocked(newpage);
263 __mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages);
H A Dshmem.c967 struct page *oldpage, *newpage; local
981 newpage = shmem_alloc_page(gfp, info, index);
982 if (!newpage)
985 page_cache_get(newpage);
986 copy_highpage(newpage, oldpage);
987 flush_dcache_page(newpage);
989 __set_page_locked(newpage);
990 SetPageUptodate(newpage);
991 SetPageSwapBacked(newpage);
992 set_page_private(newpage, swap_inde
[all...]
H A Dksm.c1956 void ksm_migrate_page(struct page *newpage, struct page *oldpage) argument
1961 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage);
1962 VM_BUG_ON_PAGE(newpage->mapping != oldpage->mapping, newpage);
1964 stable_node = page_stable_node(newpage);
1967 stable_node->kpfn = page_to_pfn(newpage);
1969 * newpage->mapping was set in advance; now we need smp_wmb()
H A Dmemcontrol.c6603 * @newpage: page to transfer the charge to
6606 * Migrate the charge from @oldpage to @newpage.
6608 * Both pages must be locked, @newpage->mapping must be set up.
6610 void mem_cgroup_migrate(struct page *oldpage, struct page *newpage, argument
6617 VM_BUG_ON_PAGE(!PageLocked(newpage), newpage);
6619 VM_BUG_ON_PAGE(!lrucare && PageLRU(newpage), newpage);
6620 VM_BUG_ON_PAGE(PageAnon(oldpage) != PageAnon(newpage), newpage); local
6622 newpage); local
[all...]

Completed in 94 milliseconds