Searched defs:pfn (Results 101 - 125 of 140) sorted by relevance

123456

/arch/powerpc/include/asm/
H A Dkvm_book3s.h63 u64 pfn; member in struct:hpte_cache
H A Drtas.h349 static inline int page_is_rtas_user_buf(unsigned long pfn) argument
351 unsigned long paddr = (pfn << PAGE_SHIFT);
357 static inline int page_is_rtas_user_buf(unsigned long pfn) { return 0;} argument
/arch/s390/kernel/
H A Dsetup.c41 #include <linux/pfn.h>
851 unsigned long start_chunk, end_chunk, pfn; local
863 pfn = max(start_chunk, start_pfn);
864 for (; pfn < end_chunk; pfn++)
865 page_set_storage_key(PFN_PHYS(pfn),
/arch/tile/mm/
H A Dinit.c93 pte_t *get_prealloc_pte(unsigned long pfn) argument
95 int node = pfn_to_nid(pfn);
96 pfn &= ~(-1UL << (NR_PA_HIGHBIT_SHIFT - PAGE_SHIFT));
98 BUG_ON(pfn >= num_l2_ptes[node]);
99 return &l2_ptes[node][pfn];
447 unsigned long address, pfn; local
497 pfn = node_start_pfn[i];
500 init_prealloc_ptes(i, end_pfn - pfn);
502 address = (unsigned long) pfn_to_kaddr(pfn);
503 while (pfn < end_pf
704 unsigned long pfn; local
970 unsigned long pfn = kaddr_to_pfn((void *)addr); local
1015 int pfn = kaddr_to_pfn((void *)addr); local
[all...]
H A Dpgtable.c90 static void set_pte_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) argument
113 /* <pfn,flags> stored as-is, to permit clearing entries */
114 set_pte(pte, pfn_pte(pfn, flags));
477 unsigned long pfn = pte_pfn(pte); local
478 if (pfn_valid(pfn)) {
480 pte = pte_set_home(pte, page_home(pfn_to_page(pfn)));
/arch/x86/kvm/
H A Dpaging_tmpl.h365 pfn_t pfn; local
373 pfn = gfn_to_pfn_atomic(vcpu->kvm, gpte_to_gfn(gpte));
374 if (mmu_invalid_pfn(pfn)) {
375 kvm_release_pfn_clean(pfn);
381 * vcpu->arch.update_pte.pfn was fetched from get_user_pages(write = 1).
385 gpte_to_gfn(gpte), pfn, true, true);
434 pfn_t pfn; local
450 pfn = pte_prefetch_gfn_to_pfn(vcpu, gfn,
452 if (mmu_invalid_pfn(pfn)) {
453 kvm_release_pfn_clean(pfn);
466 fetch(struct kvm_vcpu *vcpu, gva_t addr, struct guest_walker *gw, int user_fault, int write_fault, int hlevel, int *emulate, pfn_t pfn, bool map_writable, bool prefault) argument
579 pfn_t pfn; local
[all...]
/arch/x86/mm/
H A Dfault.c308 static bool low_pfn(unsigned long pfn) argument
310 return pfn < max_low_pfn;
H A Dinit_32.c25 #include <linux/pfn.h>
64 unsigned long pfn = pgt_buf_end++; local
67 if (pfn >= pgt_buf_top)
70 adr = __va(pfn * PAGE_SIZE);
249 unsigned long pfn; local
280 pfn = start_pfn;
281 pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
286 if (pfn >= end_pfn)
289 pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
294 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pf
433 unsigned long pfn = clamp_t(unsigned long, PFN_UP(start), local
450 unsigned long pfn, va; local
[all...]
H A Dinit_64.c27 #include <linux/pfn.h>
319 unsigned long pfn = pgt_buf_end++; local
329 if (pfn >= pgt_buf_top)
332 adr = early_memremap(pfn * PAGE_SIZE, PAGE_SIZE);
334 *phys = pfn * PAGE_SIZE;
H A Dpat.c198 u64 pfn; local
206 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) {
209 page = pfn_to_page(pfn);
225 for (pfn = (start >> PAGE_SHIFT); pfn < (end >> PAGE_SHIFT); ++pfn) {
226 page = pfn_to_page(pfn);
235 u64 pfn; local
469 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t vma_prot) argument
477 range_is_allowed(unsigned long pfn, unsigned long size) argument
483 range_is_allowed(unsigned long pfn, unsigned long size) argument
506 phys_mem_access_prot_allowed(struct file *file, unsigned long pfn, unsigned long size, pgprot_t *vma_prot) argument
692 track_pfn_vma_new(struct vm_area_struct *vma, pgprot_t *prot, unsigned long pfn, unsigned long size) argument
721 untrack_pfn_vma(struct vm_area_struct *vma, unsigned long pfn, unsigned long size) argument
[all...]
H A Dpageattr.c13 #include <linux/pfn.h>
37 unsigned long pfn; member in struct:cpa_data
254 unsigned long pfn)
263 if (pcibios_enabled && within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT))
276 * The .rodata section needs to be read-only. Using the pfn
279 if (within(pfn, __pa((unsigned long)__start_rodata) >> PAGE_SHIFT,
395 unsigned long nextpage_addr, numpages, pmask, psize, addr, pfn; local
451 pfn = pte_pfn(old_pte) + ((address & (psize - 1)) >> PAGE_SHIFT);
452 cpa->pfn = pfn;
253 static_protections(pgprot_t prot, unsigned long address, unsigned long pfn) argument
506 unsigned long pfn, pfninc = 1; local
655 unsigned long pfn = pte_pfn(old_pte); local
[all...]
/arch/x86/xen/
H A Dp2m.c55 * PFN is that we could find ourselves where pfn_to_mfn(pfn)==pfn for a
56 * non-identity pfn. To protect ourselves against we elect to set (and get) the
80 * aligned on 512^2*PAGE_SIZE (1GB) we loop on aligned 1GB PFNs from start pfn
81 * to end pfn. We reserve_brk top leaf pages if they are missing (means they
89 * on the 4MB (or 2MB depending on architecture) off the start and end pfn's.
90 * We check if the start pfn and end pfn violate that boundary check, and if
103 * The next step is to walk from the start pfn to the end pfn settin
197 p2m_top_index(unsigned long pfn) argument
203 p2m_mid_index(unsigned long pfn) argument
208 p2m_index(unsigned long pfn) argument
273 unsigned long pfn; local
343 unsigned long pfn; local
393 get_phys_to_machine(unsigned long pfn) argument
433 alloc_p2m(unsigned long pfn) argument
502 __early_alloc_p2m(unsigned long pfn) argument
547 unsigned long pfn; local
605 __set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument
645 set_phys_to_machine(unsigned long pfn, unsigned long mfn) argument
685 unsigned long pfn; local
731 unsigned long pfn; local
830 m2p_find_override_pfn(unsigned long mfn, unsigned long pfn) argument
858 unsigned long pfn, prev_pfn_type = 0, prev_pfn_level = 0; local
[all...]
/arch/ia64/kernel/
H A Defi.c890 valid_mmap_phys_addr_range (unsigned long pfn, unsigned long size) argument
892 unsigned long phys_addr = pfn << PAGE_SHIFT;
918 phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, argument
921 unsigned long phys_addr = pfn << PAGE_SHIFT;
/arch/microblaze/pci/
H A Dpci-common.c373 unsigned long pfn,
379 resource_size_t offset = ((resource_size_t)pfn) << PAGE_SHIFT;
382 if (page_is_ram(pfn))
372 pci_phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t prot) argument
/arch/mips/mm/
H A Dc-r4k.c450 unsigned long pfn; member in struct:flush_cache_page_args
458 struct page *page = pfn_to_page(fcp_args->pfn);
528 unsigned long addr, unsigned long pfn)
534 args.pfn = pfn;
527 r4k_flush_cache_page(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument
/arch/parisc/mm/
H A Dinit.c610 unsigned long pfn; local
615 for (pfn = node_start_pfn(i); pfn < node_end_pfn(i); pfn++) {
616 if (PageReserved(pfn_to_page(pfn)))
621 for (pfn = 0; pfn < max_pfn; pfn++) {
625 if (PageReserved(pfn_to_page(pfn)))
/arch/powerpc/kernel/
H A Dpci-common.c375 unsigned long pfn,
381 resource_size_t offset = ((resource_size_t)pfn) << PAGE_SHIFT;
384 if (page_is_ram(pfn))
374 pci_phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t prot) argument
/arch/powerpc/kvm/
H A Dbook3s_64_mmu_hv.c215 unsigned long pfn, i, npages; local
241 pfn = vma->vm_pgoff + ((start - vma->vm_start) >> PAGE_SHIFT);
242 /* check alignment of pfn vs. requested page size */
243 if (psize > PAGE_SIZE && (pfn & ((psize >> PAGE_SHIFT) - 1)))
268 pfn = page_to_pfn(page);
277 physp[i] = ((pfn + i) << PAGE_SHIFT) +
511 unsigned long gfn, hva, pfn; local
567 pfn = 0;
581 pfn = vma->vm_pgoff +
588 if (!pfn)
[all...]
H A Dbook3s_hv.c1183 unsigned long j, npages, pfn; local
1193 pfn = physp[j] >> PAGE_SHIFT;
1194 page = pfn_to_page(pfn);
H A De500_tlb.c346 pfn_t pfn; local
348 pfn = (pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT;
349 get_page(pfn_to_page(pfn));
357 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) |
512 pfn_t pfn)
514 ref->pfn = pfn;
525 kvm_release_pfn_dirty(ref->pfn);
527 kvm_release_pfn_clean(ref->pfn);
594 pfn_t pfn local
510 kvmppc_e500_ref_setup(struct tlbe_ref *ref, struct kvm_book3e_206_tlb_entry *gtlbe, pfn_t pfn) argument
614 unsigned long pfn, hva; local
[all...]
/arch/sparc/include/asm/
H A Dpgtable_32.h257 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot)
435 * For sparc32&64, the pfn in io_remap_pfn_range() carries <iospace> in
438 #define MK_IOSPACE_PFN(space, pfn) (pfn | (space << (BITS_PER_LONG - 4)))
439 #define GET_IOSPACE(pfn) (pfn >> (BITS_PER_LONG - 4))
440 #define GET_PFN(pfn) (pfn & 0x0fffffffUL)
446 unsigned long from, unsigned long pfn,
445 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) argument
[all...]
/arch/tile/include/asm/
H A Dpgtable.h271 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) argument
273 return hv_pte_set_pfn(prot, pfn);
363 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \
364 remap_pfn_range(vma, vaddr, pfn, size, prot)
/arch/x86/kernel/
H A De820.c17 #include <linux/pfn.h>
687 unsigned long pfn; local
689 pfn = PFN_DOWN(e820.map[0].addr + e820.map[0].size);
693 if (pfn < PFN_UP(ei->addr))
694 register_nosave_region(pfn, PFN_UP(ei->addr));
696 pfn = PFN_DOWN(ei->addr + ei->size);
698 register_nosave_region(PFN_UP(ei->addr), pfn);
700 if (pfn >= limit_pfn)
/arch/alpha/kernel/
H A Dsetup.c37 #include <linux/pfn.h>
464 page_is_ram(unsigned long pfn) argument
474 if (pfn >= cluster->start_pfn &&
475 pfn < cluster->start_pfn + cluster->numpages) {
/arch/ia64/kvm/
H A Dkvm-ia64.c1596 unsigned long pfn; local
1604 pfn = gfn_to_pfn(kvm, base_gfn + i);
1605 if (!kvm_is_mmio_pfn(pfn)) {
1607 pfn << PAGE_SHIFT,
1609 memslot->rmap[i] = (unsigned long)pfn_to_page(pfn);
1612 GPFN_PHYS_MMIO | (pfn << PAGE_SHIFT),

Completed in 472 milliseconds

123456