/arch/x86/include/asm/ |
H A D | page.h | 24 static inline void clear_user_page(void *page, unsigned long vaddr, argument 30 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, argument 36 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ 37 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
/arch/x86/kernel/ |
H A D | pci-dma.c | 133 void dma_generic_free_coherent(struct device *dev, size_t size, void *vaddr, argument 137 struct page *page = virt_to_page(vaddr); 140 free_pages((unsigned long)vaddr, get_order(size));
|
/arch/x86/mm/ |
H A D | highmem_32.c | 35 unsigned long vaddr; local 46 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 51 return (void *)vaddr; 73 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 75 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && 76 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { 83 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 91 kpte_clear_flush(kmap_pte-idx, vaddr); 97 BUG_ON(vaddr < PAGE_OFFSET); 98 BUG_ON(vaddr > 108 unsigned long idx, vaddr = (unsigned long)ptr; local [all...] |
/arch/x86/xen/ |
H A D | grant-table.c | 121 void *vaddr; local 147 vaddr = vmap(pages, nr_grant_frames, 0, PAGE_KERNEL); 148 if (!vaddr) { 160 xen_auto_xlat_grant_frames.vaddr = vaddr;
|
/arch/alpha/mm/ |
H A D | init.c | 193 unsigned long vaddr; local 205 vaddr = (unsigned long)console_remap_vm.addr; 211 crb->map[i].va = vaddr; 216 if (pmd != pmd_offset(pgd, vaddr)) { 218 pmd = pmd_offset(pgd, vaddr); 222 set_pte(pte_offset_kernel(pmd, vaddr), 225 vaddr += PAGE_SIZE;
|
/arch/arm/include/asm/xen/ |
H A D | page.h | 74 static inline xmaddr_t arbitrary_virt_to_machine(void *vaddr) argument 77 return virt_to_machine(vaddr);
|
/arch/arm/kernel/ |
H A D | tcm.c | 73 unsigned long vaddr; local 78 vaddr = gen_pool_alloc(tcm_pool, len); 79 if (!vaddr) 82 return (void *) vaddr;
|
/arch/arm/mm/ |
H A D | cache-xsc3l2.c | 100 unsigned long vaddr; local 107 vaddr = -1; /* to force the first mapping */ 113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); 114 xsc3_l2_clean_mva(vaddr); 115 xsc3_l2_inv_mva(vaddr); 123 vaddr = l2_map_va(start, vaddr); 124 xsc3_l2_inv_mva(vaddr); 132 vaddr 144 unsigned long vaddr; local 182 unsigned long vaddr; local [all...] |
/arch/arm/xen/ |
H A D | mm32.c | 67 static void xen_mm32_unmap(void *vaddr) argument 87 void *vaddr; local 95 vaddr = xen_mm32_remap_page(handle) + offset; 96 op(vaddr, len, dir); 97 xen_mm32_unmap(vaddr - offset); 106 vaddr = kmap_atomic(page); 107 op(vaddr + offset, len, dir); 108 kunmap_atomic(vaddr); 110 vaddr = kmap_high_get(page); 111 if (vaddr) { [all...] |
/arch/arm64/include/asm/ |
H A D | dma-mapping.h | 116 void *vaddr; local 118 if (dma_alloc_from_coherent(dev, size, dma_handle, &vaddr)) 119 return vaddr; 121 vaddr = ops->alloc(dev, size, dma_handle, flags, attrs); 122 debug_dma_alloc_coherent(dev, size, *dma_handle, vaddr); 123 return vaddr; 127 void *vaddr, dma_addr_t dev_addr, 132 if (dma_release_from_coherent(dev, get_order(size), vaddr)) 135 debug_dma_free_coherent(dev, size, vaddr, dev_addr); 136 ops->free(dev, size, vaddr, dev_add 126 dma_free_attrs(struct device *dev, size_t size, void *vaddr, dma_addr_t dev_addr, struct dma_attrs *attrs) argument [all...] |
/arch/blackfin/include/asm/ |
H A D | dma-mapping.h | 15 void dma_free_coherent(struct device *dev, size_t size, void *vaddr, 151 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 154 _dma_sync((dma_addr_t)vaddr, size, dir);
|
/arch/blackfin/kernel/ |
H A D | dma-mapping.c | 98 dma_free_coherent(struct device *dev, size_t size, void *vaddr, argument 101 __free_dma_pages((unsigned long)vaddr, get_pages(size));
|
/arch/cris/include/asm/ |
H A D | dma-mapping.h | 23 void *vaddr, dma_addr_t dma_handle); 156 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
/arch/frv/include/asm/ |
H A D | dma-mapping.h | 22 void dma_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle); 129 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
/arch/frv/mb93090-mb00/ |
H A D | pci-dma-nommu.c | 89 void dma_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle) argument
|
/arch/hexagon/kernel/ |
H A D | dma.c | 97 static void hexagon_free_coherent(struct device *dev, size_t size, void *vaddr, argument 100 gen_pool_free(coherent_pool, (unsigned long) vaddr, size);
|
/arch/ia64/include/asm/ |
H A D | dma-mapping.h | 99 dma_cache_sync (struct device *dev, void *vaddr, size_t size, argument
|
/arch/m68k/include/asm/ |
H A D | dma-mapping.h | 49 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
H A D | page_mm.h | 9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) 58 #define clear_user_page(addr, vaddr, page) \ 62 #define copy_user_page(to, from, vaddr, page) \ 73 static inline unsigned long ___pa(void *vaddr) argument 80 : "0" (vaddr), "i" (m68k_fixup_memoffset)); 83 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) 86 void *vaddr; local 90 : "=r" (vaddr) 92 return vaddr; [all...] |
/arch/m68k/sun3x/ |
H A D | dvma.c | 83 unsigned long vaddr, int len) 90 vaddr &= PAGE_MASK; 92 end = PAGE_ALIGN(vaddr + len); 96 kaddr, vaddr); 98 pgd = pgd_offset_k(vaddr); 104 if((pmd = pmd_alloc(&init_mm, pgd, vaddr)) == NULL) { 109 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) 110 end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK; 118 if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) { 123 if((end2 & PMD_MASK) > (vaddr 82 dvma_map_cpu(unsigned long kaddr, unsigned long vaddr, int len) argument [all...] |
/arch/metag/include/asm/ |
H A D | dma-mapping.h | 18 void *vaddr, dma_addr_t dma_handle); 20 void dma_sync_for_device(void *vaddr, size_t size, int dma_direction); 21 void dma_sync_for_cpu(void *vaddr, size_t size, int dma_direction); 178 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
/arch/metag/mm/ |
H A D | mmu-meta2.c | 18 unsigned long mmu_read_first_level_page(unsigned long vaddr) argument 25 if (is_global_space(vaddr)) 26 vaddr &= ~0x80000000; 28 offset = vaddr >> PGDIR_SHIFT; 53 unsigned long mmu_read_second_level_page(unsigned long vaddr) argument 55 return __builtin_meta2_cacherd((void *)(vaddr & PAGE_MASK));
|
/arch/microblaze/include/asm/ |
H A D | dma-mapping.h | 135 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 139 __dma_sync(virt_to_phys(vaddr), size, (int)direction);
|
/arch/microblaze/kernel/ |
H A D | dma.c | 44 void *vaddr, dma_addr_t dma_handle, 48 consistent_free(size, vaddr); 50 free_pages((unsigned long)vaddr, get_order(size)); 43 dma_direct_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
/arch/mips/dec/ |
H A D | kn01-berr.c | 83 long asid, entryhi, vaddr; local 105 vaddr = regs->regs[insn.i_format.rs] + 108 vaddr = (long)pc; 109 if (KSEGX(vaddr) == CKSEG0 || KSEGX(vaddr) == CKSEG1) 110 address = CPHYSADDR(vaddr); 115 entryhi |= vaddr & ~(PAGE_SIZE - 1); 123 offset = vaddr & (PAGE_SIZE - 1);
|