Lines Matching refs:newpage

341 		struct page *newpage, struct page *page,
389 get_page(newpage); /* add cache reference */
391 SetPageSwapCache(newpage);
392 set_page_private(newpage, page_private(page));
395 radix_tree_replace_slot(pslot, newpage);
415 __inc_zone_page_state(newpage, NR_FILE_PAGES);
418 __inc_zone_page_state(newpage, NR_SHMEM);
430 struct page *newpage, struct page *page)
458 get_page(newpage);
460 radix_tree_replace_slot(pslot, newpage);
519 void migrate_page_copy(struct page *newpage, struct page *page)
524 copy_huge_page(newpage, page);
526 copy_highpage(newpage, page);
529 SetPageError(newpage);
531 SetPageReferenced(newpage);
533 SetPageUptodate(newpage);
536 SetPageActive(newpage);
538 SetPageUnevictable(newpage);
540 SetPageChecked(newpage);
542 SetPageMappedToDisk(newpage);
554 SetPageDirty(newpage);
556 __set_page_dirty_nobuffers(newpage);
564 page_cpupid_xchg_last(newpage, cpupid);
566 mlock_migrate_page(newpage, page);
567 ksm_migrate_page(newpage, page);
580 if (PageWriteback(newpage))
581 end_page_writeback(newpage);
595 struct page *newpage, struct page *page,
602 rc = migrate_page_move_mapping(mapping, newpage, page, NULL, mode, 0);
607 migrate_page_copy(newpage, page);
619 struct page *newpage, struct page *page, enum migrate_mode mode)
625 return migrate_page(mapping, newpage, page, mode);
629 rc = migrate_page_move_mapping(mapping, newpage, page, head, mode, 0);
643 set_page_private(newpage, page_private(page));
646 get_page(newpage);
650 set_bh_page(bh, newpage, bh_offset(bh));
655 SetPagePrivate(newpage);
657 migrate_page_copy(newpage, page);
717 struct page *newpage, struct page *page, enum migrate_mode mode)
734 return migrate_page(mapping, newpage, page, mode);
748 static int move_to_new_page(struct page *newpage, struct page *page,
759 if (!trylock_page(newpage))
763 newpage->index = page->index;
764 newpage->mapping = page->mapping;
766 SetPageSwapBacked(newpage);
770 rc = migrate_page(mapping, newpage, page, mode);
779 newpage, page, mode);
781 rc = fallback_migrate_page(mapping, newpage, page, mode);
784 newpage->mapping = NULL;
786 mem_cgroup_migrate(page, newpage, false);
788 remove_migration_ptes(page, newpage);
792 unlock_page(newpage);
797 static int __unmap_and_move(struct page *page, struct page *newpage,
887 rc = balloon_page_migrate(newpage, page, mode);
917 rc = move_to_new_page(newpage, page, remap_swapcache, mode);
934 * to the newly allocated page in newpage.
942 struct page *newpage = get_new_page(page, private, &result);
944 if (!newpage)
956 rc = __unmap_and_move(page, newpage, force, mode);
978 ClearPageSwapBacked(newpage);
979 put_new_page(newpage, private);
980 } else if (unlikely(__is_movable_balloon_page(newpage))) {
982 put_page(newpage);
984 putback_lru_page(newpage);
990 *result = page_to_nid(newpage);
1586 struct page *newpage;
1588 newpage = alloc_pages_exact_node(nid,
1594 return newpage;