/arch/arm/kernel/ |
H A D | crash_dump.c | 37 void *vaddr; local 42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); 43 if (!vaddr) 47 if (copy_to_user(buf, vaddr + offset, csize)) { 48 iounmap(vaddr); 52 memcpy(buf, vaddr + offset, csize); 55 iounmap(vaddr);
|
/arch/arm64/mm/ |
H A D | copypage.c | 25 void __cpu_copy_user_page(void *kto, const void *kfrom, unsigned long vaddr) argument 32 void __cpu_clear_user_page(void *kaddr, unsigned long vaddr) argument
|
/arch/ia64/kernel/ |
H A D | crash_dump.c | 37 void *vaddr; local 41 vaddr = __va(pfn<<PAGE_SHIFT); 43 if (copy_to_user(buf, (vaddr + offset), csize)) { 47 memcpy(buf, (vaddr + offset), csize);
|
/arch/sh/boards/mach-sdk7786/ |
H A D | sram.c | 25 void __iomem *vaddr; local 53 vaddr = ioremap(phys, SZ_2K); 54 if (unlikely(!vaddr)) { 63 ret = gen_pool_add(sram_pool, (unsigned long)vaddr, SZ_2K, -1); 66 iounmap(vaddr);
|
/arch/sh/kernel/ |
H A D | crash_dump.c | 28 void *vaddr; local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 36 if (copy_to_user(buf, (vaddr + offset), csize)) { 37 iounmap(vaddr); 41 memcpy(buf, (vaddr + offset), csize); 43 iounmap(vaddr);
|
/arch/x86/kernel/ |
H A D | crash_dump_64.c | 29 void *vaddr; local 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); 35 if (!vaddr) 39 if (copy_to_user(buf, vaddr + offset, csize)) { 40 iounmap(vaddr); 44 memcpy(buf, vaddr + offset, csize); 47 iounmap(vaddr);
|
/arch/cris/arch-v32/drivers/pci/ |
H A D | dma.c | 43 void *vaddr, dma_addr_t dma_handle) 47 if (!dma_release_from_coherent(dev, order, vaddr)) 48 free_pages((unsigned long)vaddr, order); 42 dma_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle) argument
|
/arch/frv/mm/ |
H A D | cache-page.c | 25 void *vaddr; local 29 vaddr = kmap_atomic_primary(page); 31 frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE); 33 kunmap_atomic_primary(vaddr); 53 void *vaddr; local 57 vaddr = kmap_atomic_primary(page); 59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; 62 kunmap_atomic_primary(vaddr);
|
/arch/microblaze/mm/ |
H A D | highmem.c | 37 unsigned long vaddr; local 48 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); 53 local_flush_tlb_page(NULL, vaddr); 55 return (void *) vaddr; 61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 64 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { 75 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 81 pte_clear(&init_mm, vaddr, kmap_pte-idx); 82 local_flush_tlb_page(NULL, vaddr); [all...] |
/arch/mips/kernel/ |
H A D | crash_dump.c | 29 void *vaddr; local 34 vaddr = kmap_atomic_pfn(pfn); 37 memcpy(buf, (vaddr + offset), csize); 38 kunmap_atomic(vaddr); 45 copy_page(kdump_buf_page, vaddr); 46 kunmap_atomic(vaddr);
|
/arch/powerpc/mm/ |
H A D | highmem.c | 34 unsigned long vaddr; local 44 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); 49 local_flush_tlb_page(NULL, vaddr); 51 return (void*) vaddr; 57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 60 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { 72 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 78 pte_clear(&init_mm, vaddr, kmap_pte-idx); 79 local_flush_tlb_page(NULL, vaddr); [all...] |
/arch/sh/mm/ |
H A D | kmap.c | 18 #define kmap_get_fixmap_pte(vaddr) \ 19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr)) 25 unsigned long vaddr; local 28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); 29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); 35 unsigned long vaddr; local 45 vaddr 56 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; local 63 local_flush_tlb_one(get_asid(), vaddr); local [all...] |
/arch/unicore32/mm/ |
H A D | dma-swiotlb.c | 28 void *vaddr, dma_addr_t dma_addr, 31 swiotlb_free_coherent(dev, size, vaddr, dma_addr); 27 unicore_swiotlb_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_addr, struct dma_attrs *attrs) argument
|
/arch/arc/mm/ |
H A D | dma.c | 46 void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, argument
|
H A D | ioremap.c | 44 void __iomem *vaddr; local 73 vaddr = (void __iomem *)area->addr; 74 if (ioremap_page_range((unsigned long)vaddr, 75 (unsigned long)vaddr + size, paddr, prot)) { 76 vunmap((void __force *)vaddr); 79 return (void __iomem *)(off + (char __iomem *)vaddr);
|
/arch/arm/mm/ |
H A D | copypage-fa.c | 43 unsigned long vaddr, struct vm_area_struct *vma) 59 void fa_clear_user_highpage(struct page *page, unsigned long vaddr) argument 42 fa_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-feroceon.c | 71 unsigned long vaddr, struct vm_area_struct *vma) 77 flush_cache_page(vma, vaddr, page_to_pfn(from)); 83 void feroceon_clear_user_highpage(struct page *page, unsigned long vaddr) argument 70 feroceon_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-v4mc.c | 68 unsigned long vaddr, struct vm_area_struct *vma) 89 void v4_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 67 v4_mc_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-v4wb.c | 51 unsigned long vaddr, struct vm_area_struct *vma) 57 flush_cache_page(vma, vaddr, page_to_pfn(from)); 68 void v4wb_clear_user_highpage(struct page *page, unsigned long vaddr) argument 50 v4wb_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-v4wt.c | 47 unsigned long vaddr, struct vm_area_struct *vma) 63 void v4wt_clear_user_highpage(struct page *page, unsigned long vaddr) argument 46 v4wt_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-v6.c | 34 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) 49 static void v6_clear_user_highpage_nonaliasing(struct page *page, unsigned long vaddr) argument 73 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) 75 unsigned int offset = CACHE_COLOUR(vaddr); 106 static void v6_clear_user_highpage_aliasing(struct page *page, unsigned long vaddr) argument 108 unsigned long to = COPYPAGE_V6_TO + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); 33 v6_copy_user_highpage_nonaliasing(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument 72 v6_copy_user_highpage_aliasing(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-xsc3.c | 74 unsigned long vaddr, struct vm_area_struct *vma) 80 flush_cache_page(vma, vaddr, page_to_pfn(from)); 91 void xsc3_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 73 xsc3_mc_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
H A D | copypage-xscale.c | 88 unsigned long vaddr, struct vm_area_struct *vma) 110 xscale_mc_clear_user_highpage(struct page *page, unsigned long vaddr) argument 87 xscale_mc_copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument
|
/arch/arm/plat-samsung/ |
H A D | pm-debug.c | 58 unsigned long vaddr; local 60 debug_ll_addr(&paddr, &vaddr); 62 return (void __iomem *)vaddr;
|
/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 85 unsigned long vaddr, 83 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
|