/arch/s390/lib/ |
H A D | uaccess_pt.c | 43 unsigned long offset, pfn, done, size; local 62 pfn = pte_pfn(*pte); 66 to = (void *)((pfn << PAGE_SHIFT) + offset); 69 from = (void *)((pfn << PAGE_SHIFT) + offset); 92 unsigned long pfn; local 105 pfn = pte_pfn(*pte); 106 return (pfn << PAGE_SHIFT) + (uaddr & (PAGE_SIZE - 1)); 167 unsigned long offset, pfn, done, len; local 185 pfn = pte_pfn(*pte); 187 addr = (char *)(pfn << PAGE_SHIF [all...] |
/arch/sh/mm/ |
H A D | cache-sh4.c | 209 unsigned long address, pfn, phys; local 219 pfn = data->addr2; 220 phys = pfn << PAGE_SHIFT; 221 page = pfn_to_page(pfn);
|
H A D | cache.c | 138 unsigned long pfn = pte_pfn(pte); local 143 page = pfn_to_page(pfn); 144 if (pfn_valid(pfn)) { 192 unsigned long pfn) 198 data.addr2 = pfn; 191 flush_cache_page(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument
|
H A D | cache-sh5.c | 552 unsigned long eaddr, pfn; local 556 pfn = data->addr2; 558 sh64_dcache_purge_phy_page(pfn << PAGE_SHIFT);
|
/arch/tile/mm/ |
H A D | homecache.c | 253 unsigned long pfn = page_to_pfn(page); local 257 flush_remote(pfn, length, &home_mask, 0, 0, 0, NULL, NULL, 0); 258 sim_validate_lines_evicted(PFN_PHYS(pfn), pages * PAGE_SIZE);
|
/arch/um/kernel/ |
H A D | physmem.c | 9 #include <linux/pfn.h> 84 int pfn = PFN_UP(__pa(reserve_end)); local 107 bootmap_size = init_bootmem(pfn, pfn + delta);
|
/arch/x86/include/asm/xen/ |
H A D | page.h | 7 #include <linux/pfn.h> 45 extern unsigned long get_phys_to_machine(unsigned long pfn); 46 extern bool set_phys_to_machine(unsigned long pfn, unsigned long mfn); 47 extern bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); 55 extern unsigned long m2p_find_override_pfn(unsigned long mfn, unsigned long pfn); 57 static inline unsigned long pfn_to_mfn(unsigned long pfn) argument 62 return pfn; 64 mfn = get_phys_to_machine(pfn); 72 static inline int phys_to_machine_mapping_valid(unsigned long pfn) argument 77 return get_phys_to_machine(pfn) ! 82 unsigned long pfn; local 159 unsigned long pfn = mfn_to_pfn(mfn); local [all...] |
/arch/x86/kvm/ |
H A D | mmu_audit.c | 100 pfn_t pfn; local 117 pfn = gfn_to_pfn_atomic(vcpu->kvm, gfn); 119 if (is_error_pfn(pfn)) { 120 kvm_release_pfn_clean(pfn); 124 hpa = pfn << PAGE_SHIFT; 126 audit_printk(vcpu->kvm, "levels %d pfn %llx hpa %llx " 127 "ent %llxn", vcpu->arch.mmu.root_level, pfn,
|
/arch/x86/xen/ |
H A D | setup.c | 60 unsigned long pfn; local 83 for (pfn = PFN_DOWN(start); pfn <= xen_max_p2m_pfn; pfn++) 84 __set_phys_to_machine(pfn, INVALID_P2M_ENTRY); 96 unsigned long pfn; local 99 for(pfn = start; pfn < end; pfn++) { 100 unsigned long mfn = pfn_to_mfn(pfn); [all...] |
/arch/alpha/kernel/ |
H A D | core_titan.c | 466 unsigned long pfn; 522 pfn = ptes[baddr >> PAGE_SHIFT]; 523 if (!(pfn & 1)) { 528 pfn >>= 1; /* make it a true pfn */ 531 pfn << PAGE_SHIFT, 462 unsigned long pfn; local
|
H A D | core_marvel.c | 695 unsigned long pfn; 753 pfn = ptes[baddr >> PAGE_SHIFT]; 754 if (!(pfn & 1)) { 759 pfn >>= 1; /* make it a true pfn */ 762 pfn << PAGE_SHIFT, 686 unsigned long pfn; local
|
/arch/mips/include/asm/ |
H A D | pgtable.h | 81 static inline int is_zero_pfn(unsigned long pfn) argument 84 unsigned long offset_from_zero_pfn = pfn - zero_pfn; 381 extern int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot); 385 unsigned long pfn, 389 phys_t phys_addr_high = fixup_bigphys_addr(pfn << PAGE_SHIFT, size); 393 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 394 remap_pfn_range(vma, vaddr, pfn, size, prot) 406 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, 408 int phys_mem_access_prot_allowed(struct file *file, unsigned long pfn, 383 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long vaddr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
|
/arch/parisc/kernel/ |
H A D | cache.c | 507 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn) argument 512 __flush_cache_page(vma, vmaddr, page_to_phys(pfn_to_page(pfn)));
|
/arch/powerpc/mm/ |
H A D | mem.c | 81 int page_is_ram(unsigned long pfn) argument 84 return pfn < max_pfn; 86 unsigned long paddr = (pfn << PAGE_SHIFT); 96 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, argument 100 return ppc_md.phys_mem_access_prot(file, pfn, size, vma_prot); 102 if (!page_is_ram(pfn)) 339 unsigned long pfn, highmem_mapnr; local 342 for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { 598 devmem_is_allowed(unsigned long pfn) argument [all...] |
/arch/s390/include/asm/ |
H A D | page.h | 157 static inline int page_test_and_clear_dirty(unsigned long pfn, int mapped) argument 161 skey = page_get_storage_key(pfn << PAGE_SHIFT); 164 page_set_storage_key(pfn << PAGE_SHIFT, skey & ~_PAGE_CHANGED, mapped); 172 static inline int page_test_and_clear_young(unsigned long pfn) argument 174 return page_reset_referenced(pfn << PAGE_SHIFT); 182 static inline int devmem_is_allowed(unsigned long pfn) argument
|
/arch/s390/kernel/ |
H A D | crash_dump.c | 30 ssize_t copy_oldmem_page(unsigned long pfn, char *buf, argument 38 src = (pfn << PAGE_SHIFT) + offset;
|
/arch/tile/include/asm/ |
H A D | page.h | 141 #define __pfn_to_highbits(pfn) ((pfn) >> (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)) 257 static inline void *pfn_to_kaddr(unsigned long pfn) argument 259 return vbase_map[__pfn_to_highbits(pfn)] + (pfn << PAGE_SHIFT); 264 unsigned long pfn = kaddr_to_pfn(kaddr); local 265 return ((phys_addr_t)pfn << PAGE_SHIFT) + 288 static inline void *pfn_to_kaddr(unsigned long pfn) argument 290 return (void *)((pfn << PAGE_SHIFT) + PAGE_OFFSET); 315 static inline int pfn_valid(unsigned long pfn) argument [all...] |
/arch/unicore32/mm/ |
H A D | init.c | 225 int pfn_valid(unsigned long pfn) argument 227 return memblock_is_memory(pfn << PAGE_SHIFT); 316 static inline int free_area(unsigned long pfn, unsigned long end, char *s) argument 318 unsigned int pages = 0, size = (end - pfn) << (PAGE_SHIFT - 10); 320 for (; pfn < end; pfn++) { 321 struct page *page = pfn_to_page(pfn);
|
H A D | mmu.c | 81 unsigned long pfn; member in struct:map_desc 164 unsigned long end, unsigned long pfn, 169 set_pte(pte, pfn_pte(pfn, __pgprot(type->prot_pte))); 170 pfn++; 217 __pfn_to_phys((u64)md->pfn), md->virtual); 225 __pfn_to_phys((u64)md->pfn), md->virtual); 231 phys = (unsigned long)__pfn_to_phys(md->pfn); 237 __pfn_to_phys(md->pfn), addr); 368 map.pfn = __phys_to_pfn(virt_to_phys(vectors)); 378 map.pfn 163 alloc_init_pte(pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, const struct mem_type *type) argument 491 unsigned long pfn = pte_pfn(*ptep); local [all...] |
/arch/x86/kernel/ |
H A D | tboot.c | 31 #include <linux/pfn.h> 122 static int map_tboot_page(unsigned long vaddr, unsigned long pfn, argument 140 set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot));
|
/arch/x86/mm/ |
H A D | ioremap.c | 66 resource_size_t pfn, last_pfn, last_addr; local 97 for (pfn = phys_addr >> PAGE_SHIFT; pfn <= last_pfn; pfn++) { 98 int is_ram = page_is_ram(pfn); 100 if (is_ram && pfn_valid(pfn) && !PageReserved(pfn_to_page(pfn)))
|
/arch/arm/mm/ |
H A D | init.c | 288 int pfn_valid(unsigned long pfn) argument 290 return memblock_is_memory(__pfn_to_phys(pfn)); 412 static inline int free_area(unsigned long pfn, unsigned long end, char *s) argument 414 unsigned int pages = 0, size = (end - pfn) << (PAGE_SHIFT - 10); 416 for (; pfn < end; pfn++) { 417 struct page *page = pfn_to_page(pfn);
|
H A D | mmu.c | 540 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, argument 543 if (!pfn_valid(pfn)) 591 unsigned long end, unsigned long pfn, 601 set_pte_ext(pte, pfn_pte(pfn, __pgprot(type->prot_pte)), 0); 602 pfn++; 666 phys = __pfn_to_phys(md->pfn); 672 (long long)__pfn_to_phys((u64)md->pfn), addr); 685 (long long)__pfn_to_phys((u64)md->pfn), addr); 689 if ((addr | length | __pfn_to_phys(md->pfn)) & ~SUPERSECTION_MASK) { 692 (long long)__pfn_to_phys((u64)md->pfn), add 590 alloc_init_pte(pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, const struct mem_type *type) argument [all...] |
/arch/ia64/kvm/ |
H A D | vtlb.c | 118 u64 index, pfn, rid, pfn_bits; local 121 pfn = REGION_OFFSET(va) >> _REGION_PAGE_SIZE(vrr); 123 index = ((rid & 0xff) << pfn_bits)|(pfn & ((1UL << pfn_bits) - 1)); 124 *tag = ((rid >> 8) & 0xffff) | ((pfn >> pfn_bits) << 16);
|
/arch/ia64/mm/ |
H A D | init.c | 521 ia64_pfn_valid (unsigned long pfn) argument 524 struct page *pg = pfn_to_page(pfn);
|