/arch/x86/kernel/ |
H A D | tboot.c | 123 static int map_tboot_page(unsigned long vaddr, unsigned long pfn, argument 131 pgd = pgd_offset(&tboot_mm, vaddr); 132 pud = pud_alloc(&tboot_mm, pgd, vaddr); 135 pmd = pmd_alloc(&tboot_mm, pud, vaddr); 138 pte = pte_alloc_map(&tboot_mm, NULL, pmd, vaddr); 141 set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot)); 146 static int map_tboot_pages(unsigned long vaddr, unsigned long start_pfn, argument 154 for (; nr > 0; nr--, vaddr += PAGE_SIZE, start_pfn++) { 155 if (map_tboot_page(vaddr, start_pfn, PAGE_KERNEL_EXEC))
|
H A D | machine_kexec_32.c | 103 unsigned long vaddr, unsigned long paddr) 107 pgd += pgd_index(vaddr); 112 pud = pud_offset(pgd, vaddr); 113 pmd = pmd_offset(pud, vaddr); 116 pte = pte_offset_kernel(pmd, vaddr); 101 machine_kexec_page_table_set_one( pgd_t *pgd, pmd_t *pmd, pte_t *pte, unsigned long vaddr, unsigned long paddr) argument
|
/arch/m68k/sun3/ |
H A D | sun3dvma.c | 341 unsigned long vaddr; local 359 vaddr = dvma_btov(baddr); 361 if(dvma_map_cpu(kaddr, vaddr, len) < 0) { 372 return (void *)vaddr; 377 void dvma_free(void *vaddr) argument
|
/arch/powerpc/include/asm/ |
H A D | dma-mapping.h | 28 void *vaddr, dma_addr_t dma_handle, 46 extern void __dma_free_coherent(size_t size, void *vaddr); 47 extern void __dma_sync(void *vaddr, size_t size, int direction); 218 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 222 __dma_sync(vaddr, size, (int)direction);
|
/arch/blackfin/include/asm/ |
H A D | dma-mapping.h | 15 void dma_free_coherent(struct device *dev, size_t size, void *vaddr, 151 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 154 _dma_sync((dma_addr_t)vaddr, size, dir);
|
/arch/microblaze/kernel/ |
H A D | dma.c | 44 void *vaddr, dma_addr_t dma_handle, 48 consistent_free(size, vaddr); 50 free_pages((unsigned long)vaddr, get_order(size)); 43 dma_direct_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
/arch/mips/loongson/common/ |
H A D | dma-swiotlb.c | 47 void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) 51 if (dma_release_from_coherent(dev, order, vaddr)) 54 swiotlb_free_coherent(dev, size, vaddr, dma_handle); 46 loongson_dma_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
/arch/openrisc/kernel/ |
H A D | dma.c | 119 or1k_dma_free(struct device *dev, size_t size, void *vaddr, argument 122 unsigned long va = (unsigned long)vaddr; 133 free_pages_exact(vaddr, size);
|
/arch/xtensa/include/asm/ |
H A D | dma-mapping.h | 36 void *vaddr, dma_addr_t dma_handle); 167 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 170 consistent_sync(vaddr, size, direction);
|
/arch/arm/include/asm/xen/ |
H A D | page.h | 74 static inline xmaddr_t arbitrary_virt_to_machine(void *vaddr) argument 77 return virt_to_machine(vaddr);
|
/arch/c6x/mm/ |
H A D | dma-coherent.c | 102 void dma_free_coherent(struct device *dev, size_t size, void *vaddr, argument 112 __free_dma_pages(virt_to_phys(vaddr), order);
|
/arch/m68k/mm/ |
H A D | sun3kmap.c | 25 extern void print_pte_vaddr(unsigned long vaddr); 28 extern void mmu_emu_map_pmeg (int context, int vaddr);
|
/arch/metag/include/asm/ |
H A D | page.h | 68 #define clear_user_page(page, vaddr, pg) clear_page(page) 69 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from)
|
/arch/microblaze/include/asm/ |
H A D | cacheflush.h | 106 struct page *page, unsigned long vaddr, 118 struct page *page, unsigned long vaddr, 105 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument 117 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
|
/arch/mn10300/include/asm/ |
H A D | page.h | 32 #define clear_user_page(addr, vaddr, page) clear_page(addr) 33 #define copy_user_page(vto, vfrom, vaddr, to) copy_page(vto, vfrom)
|
/arch/powerpc/kernel/ |
H A D | dma-iommu.c | 29 void *vaddr, dma_addr_t dma_handle, 32 iommu_free_coherent(get_iommu_table_base(dev), size, vaddr, dma_handle); local 28 dma_iommu_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
H A D | iommu.c | 460 unsigned long vaddr, npages, entry, slen; local 469 vaddr = (unsigned long) sg_virt(s); 470 npages = iommu_num_pages(vaddr, slen, IOMMU_PAGE_SIZE(tbl)); 473 (vaddr & ~PAGE_MASK) == 0) 478 DBG(" - vaddr: %lx, size: %lx\n", vaddr, slen); 484 "vaddr %lx npages %lu\n", tbl, vaddr, 499 vaddr & IOMMU_PAGE_MASK(tbl), 559 unsigned long vaddr, npage local 762 void *vaddr; local 860 iommu_free_coherent(struct iommu_table *tbl, size_t size, void *vaddr, dma_addr_t dma_handle) argument [all...] |
H A D | uprobes.c | 47 * @addr: vaddr to probe. 127 regs->nip = utask->vaddr + MAX_UINSN_BYTES; 172 instruction_pointer_set(regs, utask->vaddr);
|
H A D | fadump.c | 549 void *vaddr; local 554 vaddr = (void *)__get_free_pages(GFP_KERNEL|__GFP_ZERO, order); 555 if (!vaddr) 559 page = virt_to_page(vaddr); 562 return vaddr; 565 static void fadump_cpu_notes_buf_free(unsigned long vaddr, unsigned long size) argument 572 page = virt_to_page(vaddr); 597 void *vaddr; local 607 vaddr = __va(addr); 609 reg_header = vaddr; 943 void *vaddr; local [all...] |
/arch/sparc/include/asm/ |
H A D | page_32.h | 21 #define clear_user_page(addr, vaddr, page) \ 25 #define copy_user_page(to, from, vaddr, page) \
|
/arch/sparc/kernel/ |
H A D | dtlb_prot.S | 24 mov TLB_TAG_ACCESS, %g4 ! For reload of vaddr
|
/arch/sparc/mm/ |
H A D | tsb.c | 19 static inline unsigned long tsb_hash(unsigned long vaddr, unsigned long hash_shift, unsigned long nentries) argument 21 vaddr >>= hash_shift; 22 return vaddr & (nentries - 1); 25 static inline int tag_compare(unsigned long tag, unsigned long vaddr) argument 27 return (tag == (vaddr >> 22)); 97 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr) argument 107 __flush_tsb_one_entry(base, vaddr, PAGE_SHIFT, nentries); 115 __flush_tsb_one_entry(base, vaddr, REAL_HPAGE_SHIFT, nentries);
|
/arch/tile/include/asm/ |
H A D | cacheflush.h | 65 struct page *page, unsigned long vaddr, 75 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 64 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
|
/arch/unicore32/include/asm/ |
H A D | dma-mapping.h | 110 static inline void dma_cache_sync(struct device *dev, void *vaddr, argument 113 unsigned long start = (unsigned long)vaddr;
|
/arch/x86/xen/ |
H A D | mmu.c | 126 unsigned long arbitrary_virt_to_mfn(void *vaddr) argument 128 xmaddr_t maddr = arbitrary_virt_to_machine(vaddr); 133 xmaddr_t arbitrary_virt_to_machine(void *vaddr) argument 135 unsigned long address = (unsigned long)vaddr; 141 * if the PFN is in the linear mapped vaddr range, we can just use 144 if (virt_addr_valid(vaddr)) 145 return virt_to_machine(vaddr); 156 void make_lowmem_page_readonly(void *vaddr) argument 159 unsigned long address = (unsigned long)vaddr; 164 return; /* vaddr missin 172 make_lowmem_page_readwrite(void *vaddr) argument 288 set_pte_mfn(unsigned long vaddr, unsigned long mfn, pgprot_t flags) argument 1142 xen_cleanhighmap(unsigned long vaddr, unsigned long vaddr_end) argument 2067 unsigned long vaddr = __fix_to_virt(idx); local 2201 xen_zap_pfn_range(unsigned long vaddr, unsigned int order, unsigned long *in_frames, unsigned long *out_frames) argument 2229 xen_remap_exchanged_ptes(unsigned long vaddr, int order, unsigned long *mfns, unsigned long first_mfn) argument [all...] |