/arch/powerpc/include/asm/ |
H A D | kvm_book3s.h | 63 u64 pfn; member in struct:hpte_cache
|
H A D | rtas.h | 349 static inline int page_is_rtas_user_buf(unsigned long pfn) argument 351 unsigned long paddr = (pfn << PAGE_SHIFT); 357 static inline int page_is_rtas_user_buf(unsigned long pfn) { return 0;} argument
|
/arch/s390/kernel/ |
H A D | setup.c | 41 #include <linux/pfn.h> 851 unsigned long start_chunk, end_chunk, pfn; local 863 pfn = max(start_chunk, start_pfn); 864 for (; pfn < end_chunk; pfn++) 865 page_set_storage_key(PFN_PHYS(pfn),
|
/arch/tile/mm/ |
H A D | init.c | 93 pte_t *get_prealloc_pte(unsigned long pfn) argument 95 int node = pfn_to_nid(pfn); 96 pfn &= ~(-1UL << (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT)); 98 BUG_ON(pfn >= num_l2_ptes[node]); 99 return &l2_ptes[node][pfn]; 447 unsigned long address, pfn; local 497 pfn = node_start_pfn[i]; 500 init_prealloc_ptes(i, end_pfn - pfn); 502 address = (unsigned long) pfn_to_kaddr(pfn); 503 while (pfn < end_pf 704 unsigned long pfn; local 970 unsigned long pfn = kaddr_to_pfn((void *)addr); local 1015 int pfn = kaddr_to_pfn((void *)addr); local [all...] |
H A D | pgtable.c | 90 static void set_pte_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) argument 113 /* <pfn,flags> stored as-is, to permit clearing entries */ 114 set_pte(pte, pfn_pte(pfn, flags)); 477 unsigned long pfn = pte_pfn(pte); local 478 if (pfn_valid(pfn)) { 480 pte = pte_set_home(pte, page_home(pfn_to_page(pfn)));
|
/arch/x86/kvm/ |
H A D | paging_tmpl.h | 365 pfn_t pfn; local 373 pfn = gfn_to_pfn_atomic(vcpu->kvm, gpte_to_gfn(gpte)); 374 if (mmu_invalid_pfn(pfn)) { 375 kvm_release_pfn_clean(pfn); 381 * vcpu->arch.update_pte.pfn was fetched from get_user_pages(write = 1). 385 gpte_to_gfn(gpte), pfn, true, true); 434 pfn_t pfn; local 450 pfn = pte_prefetch_gfn_to_pfn(vcpu, gfn, 452 if (mmu_invalid_pfn(pfn)) { 453 kvm_release_pfn_clean(pfn); 466 fetch(struct kvm_vcpu *vcpu, gva_t addr, struct guest_walker *gw, int user_fault, int write_fault, int hlevel, int *emulate, pfn_t pfn, bool map_writable, bool prefault) argument 579 pfn_t pfn; local [all...] |
/arch/x86/mm/ |
H A D | fault.c | 308 static bool low_pfn(unsigned long pfn) argument 310 return pfn < max_low_pfn;
|
H A D | init_32.c | 25 #include <linux/pfn.h> 64 unsigned long pfn = pgt_buf_end++; local 67 if (pfn >= pgt_buf_top) 70 adr = __va(pfn * PAGE_SIZE); 249 unsigned long pfn; local 280 pfn = start_pfn; 281 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); 286 if (pfn >= end_pfn) 289 pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET); 294 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pf 433 unsigned long pfn = clamp_t(unsigned long, PFN_UP(start), local 450 unsigned long pfn, va; local [all...] |
H A D | init_64.c | 27 #include <linux/pfn.h> 319 unsigned long pfn = pgt_buf_end++; local 329 if (pfn >= pgt_buf_top) 332 adr = early_memremap(pfn * PAGE_SIZE, PAGE_SIZE); 334 *phys = pfn * PAGE_SIZE;
|
H A D | pat.c | 198 u64 pfn; local 206 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { 209 page = pfn_to_page(pfn); 225 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) { 226 page = pfn_to_page(pfn); 235 u64 pfn; local 469 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t vma_prot) argument 477 range_is_allowed(unsigned long pfn, unsigned long size) argument 483 range_is_allowed(unsigned long pfn, unsigned long size) argument 506 phys_mem_access_prot_allowed(struct file *file, unsigned long pfn, unsigned long size, pgprot_t *vma_prot) argument 692 track_pfn_vma_new(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long size) argument 721 untrack_pfn_vma(struct vm_area_struct *vma, unsigned long pfn, unsigned long size) argument [all...] |
H A D | pageattr.c | 13 #include <linux/pfn.h> 37 unsigned long pfn; member in struct:cpa_data 254 unsigned long pfn) 263 if (pcibios_enabled && within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) 276 * The .rodata section needs to be read-only. Using the pfn 279 if (within(pfn, __pa((unsigned long)__start_rodata) >> PAGE_SHIFT, 395 unsigned long nextpage_addr, numpages, pmask, psize, addr, pfn; local 451 pfn = pte_pfn(old_pte) + ((address & (psize - 1)) >> PAGE_SHIFT); 452 cpa->pfn = pfn; 253 static_protections(pgprot_t prot, unsigned long address, unsigned long pfn) argument 506 unsigned long pfn, pfninc = 1; local 655 unsigned long pfn = pte_pfn(old_pte); local [all...] |
/arch/x86/xen/ |
H A D | p2m.c | 55 * PFN is that we could find ourselves where pfn_to_mfn(pfn)==pfn for a 56 * non-identity pfn. To protect ourselves against we elect to set (and get) the 80 * aligned on 512^2*PAGE_SIZE (1GB) we loop on aligned 1GB PFNs from start pfn 81 * to end pfn. We reserve_brk top leaf pages if they are missing (means they 89 * on the 4MB (or 2MB depending on architecture) off the start and end pfn's. 90 * We check if the start pfn and end pfn violate that boundary check, and if 103 * The next step is to walk from the start pfn to the end pfn settin 197 p2m_top_index(unsigned long pfn) argument 203 p2m_mid_index(unsigned long pfn) argument 208 p2m_index(unsigned long pfn) argument 273 unsigned long pfn; local 343 unsigned long pfn; local 393 get_phys_to_machine(unsigned long pfn) argument 433 alloc_p2m(unsigned long pfn) argument 502 __early_alloc_p2m(unsigned long pfn) argument 547 unsigned long pfn; local 605 __set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument 645 set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument 685 unsigned long pfn; local 731 unsigned long pfn; local 830 m2p_find_override_pfn(unsigned long mfn, unsigned long pfn) argument 858 unsigned long pfn, prev_pfn_type = 0, prev_pfn_level = 0; local [all...] |
/arch/ia64/kernel/ |
H A D | efi.c | 890 valid_mmap_phys_addr_range (unsigned long pfn, unsigned long size) argument 892 unsigned long phys_addr = pfn << PAGE_SHIFT; 918 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, argument 921 unsigned long phys_addr = pfn << PAGE_SHIFT;
|
/arch/microblaze/pci/ |
H A D | pci-common.c | 373 unsigned long pfn, 379 resource_size_t offset = ((resource_size_t)pfn) << PAGE_SHIFT; 382 if (page_is_ram(pfn)) 372 pci_phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t prot) argument
|
/arch/mips/mm/ |
H A D | c-r4k.c | 450 unsigned long pfn; member in struct:flush_cache_page_args 458 struct page *page = pfn_to_page(fcp_args->pfn); 528 unsigned long addr, unsigned long pfn) 534 args.pfn = pfn; 527 r4k_flush_cache_page(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument
|
/arch/parisc/mm/ |
H A D | init.c | 610 unsigned long pfn; local 615 for (pfn = node_start_pfn(i); pfn < node_end_pfn(i); pfn++) { 616 if (PageReserved(pfn_to_page(pfn))) 621 for (pfn = 0; pfn < max_pfn; pfn++) { 625 if (PageReserved(pfn_to_page(pfn)))
|
/arch/powerpc/kernel/ |
H A D | pci-common.c | 375 unsigned long pfn, 381 resource_size_t offset = ((resource_size_t)pfn) << PAGE_SHIFT; 384 if (page_is_ram(pfn)) 374 pci_phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t prot) argument
|
/arch/powerpc/kvm/ |
H A D | book3s_64_mmu_hv.c | 215 unsigned long pfn, i, npages; local 241 pfn = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT); 242 /* check alignment of pfn vs. requested page size */ 243 if (psize > PAGE_SIZE && (pfn & ((psize >> PAGE_SHIFT) - 1))) 268 pfn = page_to_pfn(page); 277 physp[i] = ((pfn + i) << PAGE_SHIFT) + 511 unsigned long gfn, hva, pfn; local 567 pfn = 0; 581 pfn = vma->vm_pgoff + 588 if (!pfn) [all...] |
H A D | book3s_hv.c | 1183 unsigned long j, npages, pfn; local 1193 pfn = physp[j] >> PAGE_SHIFT; 1194 page = pfn_to_page(pfn);
|
H A D | e500_tlb.c | 346 pfn_t pfn; local 348 pfn = (pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; 349 get_page(pfn_to_page(pfn)); 357 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) | 512 pfn_t pfn) 514 ref->pfn = pfn; 525 kvm_release_pfn_dirty(ref->pfn); 527 kvm_release_pfn_clean(ref->pfn); 594 pfn_t pfn local 510 kvmppc_e500_ref_setup(struct tlbe_ref *ref, struct kvm_book3e_206_tlb_entry *gtlbe, pfn_t pfn) argument 614 unsigned long pfn, hva; local [all...] |
/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 257 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) 435 * For sparc32&64, the pfn in io_remap_pfn_range() carries <iospace> in 438 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4))) 439 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4)) 440 #define GET_PFN(pfn) (pfn & 0x0fffffffUL) 446 unsigned long from, unsigned long pfn, 445 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument [all...] |
/arch/tile/include/asm/ |
H A D | pgtable.h | 271 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) argument 273 return hv_pte_set_pfn(prot, pfn); 363 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ 364 remap_pfn_range(vma, vaddr, pfn, size, prot)
|
/arch/x86/kernel/ |
H A D | e820.c | 17 #include <linux/pfn.h> 687 unsigned long pfn; local 689 pfn = PFN_DOWN(e820.map[0].addr + e820.map[0].size); 693 if (pfn < PFN_UP(ei->addr)) 694 register_nosave_region(pfn, PFN_UP(ei->addr)); 696 pfn = PFN_DOWN(ei->addr + ei->size); 698 register_nosave_region(PFN_UP(ei->addr), pfn); 700 if (pfn >= limit_pfn)
|
/arch/alpha/kernel/ |
H A D | setup.c | 37 #include <linux/pfn.h> 464 page_is_ram(unsigned long pfn) argument 474 if (pfn >= cluster->start_pfn && 475 pfn < cluster->start_pfn + cluster->numpages) {
|
/arch/ia64/kvm/ |
H A D | kvm-ia64.c | 1596 unsigned long pfn; local 1604 pfn = gfn_to_pfn(kvm, base_gfn + i); 1605 if (!kvm_is_mmio_pfn(pfn)) { 1607 pfn << PAGE_SHIFT, 1609 memslot->rmap[i] = (unsigned long)pfn_to_page(pfn); 1612 GPFN_PHYS_MMIO | (pfn << PAGE_SHIFT),
|