/arch/powerpc/platforms/ps3/ |
H A D | system-bus.c | 555 static void ps3_free_coherent(struct device *_dev, size_t size, void *vaddr, argument 561 free_pages((unsigned long)vaddr, get_order(size));
|
/arch/s390/include/asm/ |
H A D | dma-mapping.h | 23 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
H A D | idals.h | 33 idal_is_needed(void *vaddr, unsigned int length) argument 36 return ((__pa(vaddr) + length - 1) >> 31) != 0; 46 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) argument 48 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + 56 void *vaddr, unsigned int length) 61 paddr = __pa(vaddr); 78 set_normalized_cda(struct ccw1 * ccw, void *vaddr) argument 86 nridaws = idal_nr_words(vaddr, ccw->count); 92 idal_create_words(idal, vaddr, ccw->count); 94 vaddr 55 idal_create_words(unsigned long *idaws, void *vaddr, unsigned int length) argument [all...] |
H A D | page.h | 68 #define clear_user_page(page, vaddr, pg) clear_page(page) 69 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) 71 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ 72 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
H A D | pgtable.h | 56 #define ZERO_PAGE(vaddr) \ 58 (((unsigned long)(vaddr)) &zero_page_mask))))
|
/arch/s390/kernel/ |
H A D | uprobes.c | 62 regs->psw.addr += utask->vaddr - utask->xol_vaddr; 66 regs->gprs[reg] += utask->vaddr - utask->xol_vaddr; 72 regs->psw.addr = utask->vaddr + ilen; 108 regs->psw.addr = current->utask->vaddr;
|
/arch/s390/kvm/ |
H A D | gaccess.c | 350 union vaddress vaddr = {.addr = gva}; local 368 if (vaddr.rfx01 > asce.tl) 370 ptr += vaddr.rfx * 8; 373 if (vaddr.rfx) 375 if (vaddr.rsx01 > asce.tl) 377 ptr += vaddr.rsx * 8; 380 if (vaddr.rfx || vaddr.rsx) 382 if (vaddr.rtx01 > asce.tl) 384 ptr += vaddr [all...] |
/arch/score/include/asm/ |
H A D | cacheflush.h | 38 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 41 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 45 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
H A D | fixmap.h | 77 static inline unsigned long virt_to_fix(const unsigned long vaddr) argument 79 return __virt_to_fix(vaddr);
|
H A D | page.h | 36 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) 37 #define copy_user_page(vto, vfrom, vaddr, topg) \ 67 #define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) 70 #define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) 83 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr)))
|
H A D | pgtable.h | 178 #define ZERO_PAGE(vaddr) \ 180 (((unsigned long)(vaddr)) & zero_page_mask))))
|
H A D | tlbflush.h | 23 extern void local_flush_tlb_one(unsigned long vaddr); 32 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr)
|
/arch/sh/boards/mach-sdk7786/ |
H A D | sram.c | 25 void __iomem *vaddr; local 53 vaddr = ioremap(phys, SZ_2K); 54 if (unlikely(!vaddr)) { 63 ret = gen_pool_add(sram_pool, (unsigned long)vaddr, SZ_2K, -1); 66 iounmap(vaddr);
|
/arch/sh/include/asm/ |
H A D | cacheflush.h | 20 * - flush_cache_sigtramp(vaddr) flushes the signal trampoline 82 struct page *page, unsigned long vaddr, void *dst, const void *src, 86 struct page *page, unsigned long vaddr, void *dst, const void *src,
|
H A D | dma-mapping.h | 39 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, 79 void *vaddr, dma_addr_t dma_handle, 84 if (dma_release_from_coherent(dev, get_order(size), vaddr)) 87 debug_dma_free_coherent(dev, size, vaddr, dma_handle); 89 ops->free(dev, size, vaddr, dma_handle, attrs); 97 void *vaddr, dma_addr_t dma_handle, 78 dma_free_attrs(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
H A D | page.h | 67 unsigned long vaddr, struct vm_area_struct *vma); 69 extern void clear_user_highpage(struct page *page, unsigned long vaddr);
|
H A D | pgtable.h | 32 #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page))
|
/arch/sh/include/cpu-sh4/cpu/ |
H A D | sq.h | 33 void sq_unmap(unsigned long vaddr);
|
/arch/sh/kernel/cpu/sh4/ |
H A D | sq.c | 203 * @vaddr: Pre-allocated Store Queue mapping. 209 void sq_unmap(unsigned long vaddr) argument 215 if (map->sq_addr == vaddr) 220 __func__, vaddr);
|
/arch/sh/kernel/ |
H A D | crash_dump.c | 28 void *vaddr; local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 36 if (copy_to_user(buf, (vaddr + offset), csize)) { 37 iounmap(vaddr); 41 memcpy(buf, (vaddr + offset), csize); 43 iounmap(vaddr);
|
H A D | io_trapped.c | 169 unsigned long vaddr = (unsigned long)tiop->virt_base; local 176 if (address < (vaddr + len)) 177 return res->start + (address - vaddr); 178 vaddr += len;
|
H A D | smp.c | 460 void flush_tlb_one(unsigned long asid, unsigned long vaddr) argument 465 fd.addr2 = vaddr; 468 local_flush_tlb_one(asid, vaddr);
|
/arch/sh/mm/ |
H A D | cache-sh4.c | 215 void *vaddr; local 236 vaddr = NULL; 246 vaddr = kmap_coherent(page, address); 248 vaddr = kmap_atomic(page); 250 address = (unsigned long)vaddr; 259 if (vaddr) { 261 kunmap_coherent(vaddr); 263 kunmap_atomic(vaddr);
|
H A D | cache-sh5.c | 597 static void sh5_flush_cache_sigtramp(void *vaddr) argument 599 unsigned long end = (unsigned long)vaddr + L1_CACHE_BYTES; 601 __flush_wback_region(vaddr, L1_CACHE_BYTES); 603 sh64_icache_inv_current_user_range((unsigned long)vaddr, end);
|
H A D | cache.c | 59 unsigned long vaddr, void *dst, const void *src, 64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); 74 flush_cache_page(vma, vaddr, page_to_pfn(page)); 78 unsigned long vaddr, void *dst, const void *src, 83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); 94 unsigned long vaddr, struct vm_area_struct *vma) 102 vfrom = kmap_coherent(from, vaddr); 111 if (pages_do_alias((unsigned long)vto, vaddr 58 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 77 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 93 copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument 121 clear_user_highpage(struct page *page, unsigned long vaddr) argument [all...] |