/mm/ |
H A D | fremap.c | 26 unsigned long addr, pte_t *ptep) 28 pte_t pte = *ptep; 34 pte = ptep_clear_flush(vma, addr, ptep); 47 pte_clear_not_present_full(mm, addr, ptep, 0); 25 zap_pte(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, pte_t *ptep) argument
|
H A D | pgtable-generic.c | 24 unsigned long address, pte_t *ptep, 27 int changed = !pte_same(*ptep, entry); 29 set_pte_at(vma->vm_mm, address, ptep, entry); 58 unsigned long address, pte_t *ptep) 61 young = ptep_test_and_clear_young(vma, address, ptep); 86 pte_t *ptep) 89 pte = ptep_get_and_clear((vma)->vm_mm, address, ptep); 23 ptep_set_access_flags(struct vm_area_struct *vma, unsigned long address, pte_t *ptep, pte_t entry, int dirty) argument 57 ptep_clear_flush_young(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument 85 ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument
|
H A D | mincore.c | 32 pte_t *ptep; local 37 ptep = huge_pte_offset(current->mm, 39 present = ptep && !huge_pte_none(huge_ptep_get(ptep)); 114 pte_t *ptep; local 116 ptep = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); 118 pte_t pte = *ptep; 146 } while (ptep++, addr = next, addr != end); 147 pte_unmap_unlock(ptep - 1, ptl);
|
H A D | migrate.c | 96 pte_t *ptep, pte; local 100 ptep = huge_pte_offset(mm, addr); 101 if (!ptep) 119 ptep = pte_offset_map(pmd, addr); 130 pte = *ptep; 149 set_pte_at(mm, addr, ptep, pte); 162 update_mmu_cache(vma, addr, ptep); 164 pte_unmap_unlock(ptep, ptl); 186 pte_t *ptep, pte; local 191 ptep [all...] |
H A D | ksm.c | 705 pte_t *ptep; local 715 ptep = page_check_address(page, mm, addr, &ptl, 0); 716 if (!ptep) 719 if (pte_write(*ptep) || pte_dirty(*ptep)) { 733 entry = ptep_clear_flush(vma, addr, ptep); 739 set_pte_at(mm, addr, ptep, entry); 745 set_pte_at_notify(mm, addr, ptep, entry); 747 *orig_pte = *ptep; 751 pte_unmap_unlock(ptep, pt 772 pte_t *ptep; local [all...] |
H A D | memory.c | 287 * Must perform the equivalent to __free_pte(pte_get_and_clear(ptep)), while 1449 pte_t *ptep, pte; local 1508 ptep = pte_offset_map_lock(mm, pmd, address, &ptl); 1510 pte = *ptep; 1560 pte_unmap_unlock(ptep, ptl); 1565 pte_unmap_unlock(ptep, ptl); 1569 pte_unmap_unlock(ptep, ptl); 3652 pte_t *ptep; local 3671 ptep = pte_offset_map_lock(mm, pmd, address, ptlp); 3672 if (!ptep) 3710 pte_t *ptep; local 3730 pte_t *ptep, pte; local [all...] |
H A D | vmalloc.c | 222 pte_t *ptep, pte; local 224 ptep = pte_offset_map(pmd, addr); 225 pte = *ptep; 228 pte_unmap(ptep);
|
H A D | hugetlb.c | 2131 unsigned long address, pte_t *ptep) 2135 entry = pte_mkwrite(pte_mkdirty(huge_ptep_get(ptep))); 2136 if (huge_ptep_set_access_flags(vma, address, ptep, entry, 1)) 2137 update_mmu_cache(vma, address, ptep); 2216 pte_t *ptep; local 2237 ptep = huge_pte_offset(mm, address); 2238 if (!ptep) 2241 if (huge_pmd_unshare(mm, &address, ptep)) 2250 pte = huge_ptep_get(ptep); 2265 pte = huge_ptep_get_and_clear(mm, address, ptep); 2130 set_huge_ptep_writable(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) argument 2355 hugetlb_cow(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, pte_t pte, struct page *pagecache_page) argument 2506 hugetlb_no_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *ptep, unsigned int flags) argument 2637 pte_t *ptep; local 2838 pte_t *ptep; local [all...] |