/arch/powerpc/platforms/ps3/ |
H A D | system-bus.c | 555 static void ps3_free_coherent(struct device *_dev, size_t size, void *vaddr, argument 561 free_pages((unsigned long)vaddr, get_order(size));
|
/arch/s390/include/asm/ |
H A D | dma-mapping.h | 23 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
H A D | idals.h | 33 idal_is_needed(void *vaddr, unsigned int length) argument 36 return ((__pa(vaddr) + length - 1) >> 31) != 0; 46 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) argument 48 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + 56 void *vaddr, unsigned int length) 61 paddr = __pa(vaddr); 78 set_normalized_cda(struct ccw1 * ccw, void *vaddr) argument 86 nridaws = idal_nr_words(vaddr, ccw->count); 92 idal_create_words(idal, vaddr, ccw->count); 94 vaddr 55 idal_create_words(unsigned long *idaws, void *vaddr, unsigned int length) argument [all...] |
/arch/s390/kvm/ |
H A D | gaccess.c | 350 union vaddress vaddr = {.addr = gva}; local 368 if (vaddr.rfx01 > asce.tl) 370 ptr += vaddr.rfx * 8; 373 if (vaddr.rfx) 375 if (vaddr.rsx01 > asce.tl) 377 ptr += vaddr.rsx * 8; 380 if (vaddr.rfx || vaddr.rsx) 382 if (vaddr.rtx01 > asce.tl) 384 ptr += vaddr [all...] |
/arch/score/include/asm/ |
H A D | fixmap.h | 77 static inline unsigned long virt_to_fix(const unsigned long vaddr) argument 79 return __virt_to_fix(vaddr);
|
/arch/sh/boards/mach-sdk7786/ |
H A D | sram.c | 25 void __iomem *vaddr; local 53 vaddr = ioremap(phys, SZ_2K); 54 if (unlikely(!vaddr)) { 63 ret = gen_pool_add(sram_pool, (unsigned long)vaddr, SZ_2K, -1); 66 iounmap(vaddr);
|
/arch/sh/include/asm/ |
H A D | dma-mapping.h | 39 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, 79 void *vaddr, dma_addr_t dma_handle, 84 if (dma_release_from_coherent(dev, get_order(size), vaddr)) 87 debug_dma_free_coherent(dev, size, vaddr, dma_handle); 89 ops->free(dev, size, vaddr, dma_handle, attrs); 97 void *vaddr, dma_addr_t dma_handle, 78 dma_free_attrs(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
/arch/sh/kernel/cpu/sh4/ |
H A D | sq.c | 203 * @vaddr: Pre-allocated Store Queue mapping. 209 void sq_unmap(unsigned long vaddr) argument 215 if (map->sq_addr == vaddr) 220 __func__, vaddr);
|
/arch/sh/kernel/ |
H A D | crash_dump.c | 28 void *vaddr; local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 36 if (copy_to_user(buf, (vaddr + offset), csize)) { 37 iounmap(vaddr); 41 memcpy(buf, (vaddr + offset), csize); 43 iounmap(vaddr);
|
H A D | io_trapped.c | 169 unsigned long vaddr = (unsigned long)tiop->virt_base; local 176 if (address < (vaddr + len)) 177 return res->start + (address - vaddr); 178 vaddr += len;
|
H A D | smp.c | 460 void flush_tlb_one(unsigned long asid, unsigned long vaddr) argument 465 fd.addr2 = vaddr; 468 local_flush_tlb_one(asid, vaddr);
|
/arch/sh/mm/ |
H A D | cache-sh4.c | 215 void *vaddr; local 236 vaddr = NULL; 246 vaddr = kmap_coherent(page, address); 248 vaddr = kmap_atomic(page); 250 address = (unsigned long)vaddr; 259 if (vaddr) { 261 kunmap_coherent(vaddr); 263 kunmap_atomic(vaddr);
|
H A D | cache-sh5.c | 597 static void sh5_flush_cache_sigtramp(void *vaddr) argument 599 unsigned long end = (unsigned long)vaddr + L1_CACHE_BYTES; 601 __flush_wback_region(vaddr, L1_CACHE_BYTES); 603 sh64_icache_inv_current_user_range((unsigned long)vaddr, end);
|
H A D | cache.c | 59 unsigned long vaddr, void *dst, const void *src, 64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); 74 flush_cache_page(vma, vaddr, page_to_pfn(page)); 78 unsigned long vaddr, void *dst, const void *src, 83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); 94 unsigned long vaddr, struct vm_area_struct *vma) 102 vfrom = kmap_coherent(from, vaddr); 111 if (pages_do_alias((unsigned long)vto, vaddr 58 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 77 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 93 copy_user_highpage(struct page *to, struct page *from, unsigned long vaddr, struct vm_area_struct *vma) argument 121 clear_user_highpage(struct page *page, unsigned long vaddr) argument [all...] |
H A D | consistent.c | 68 void *vaddr, dma_addr_t dma_handle, 78 iounmap(vaddr); 81 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 87 (void *)CAC_ADDR((unsigned long)vaddr) : vaddr; 67 dma_generic_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
H A D | init.c | 153 unsigned long vaddr, pte_t *lastpte) 166 unsigned long vaddr; local 168 vaddr = start; 169 i = __pgd_offset(vaddr); 170 j = __pud_offset(vaddr); 171 k = __pmd_offset(vaddr); 174 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { 176 for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { 181 for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) { 183 pmd, vaddr, pt 152 page_table_kmap_check(pte_t *pte, pmd_t *pmd, unsigned long vaddr, pte_t *lastpte) argument 324 unsigned long vaddr, end; local [all...] |
H A D | ioremap.c | 108 unsigned long vaddr = (unsigned long __force)addr; local 114 if (iomapping_nontranslatable(vaddr)) 129 p = remove_vm_area((void *)(vaddr & PAGE_MASK));
|
H A D | kmap.c | 18 #define kmap_get_fixmap_pte(vaddr) \ 19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr)) 25 unsigned long vaddr; local 28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); 29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); 35 unsigned long vaddr; local 45 vaddr 56 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; local 63 local_flush_tlb_one(get_asid(), vaddr); local [all...] |
H A D | pmb.c | 137 static bool pmb_mapping_exists(unsigned long vaddr, phys_addr_t phys, argument 156 if ((vaddr < pmbe->vpn) || (vaddr >= (pmbe->vpn + pmbe->size))) 335 int pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, argument 345 if (!pmb_addr_valid(vaddr, size)) 347 if (pmb_mapping_exists(vaddr, phys, size)) 350 orig_addr = vaddr; 353 flush_tlb_kernel_range(vaddr, vaddr + size); 363 pmbe = pmb_alloc(vaddr, phy 413 unsigned long vaddr; local 464 unsigned long vaddr = (unsigned long __force)addr; local [all...] |
/arch/sparc/include/asm/ |
H A D | dma-mapping.h | 15 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
H A D | floppy_64.h | 202 unsigned char *vaddr = pdma_vaddr; local 209 pdma_vaddr = vaddr; 214 pdma_vaddr = vaddr; 221 *vaddr++ = readb(stat + 1); 223 unsigned char data = *vaddr++; 231 pdma_vaddr = vaddr;
|
H A D | hvtramp.h | 9 __u64 vaddr; member in struct:hvtramp_mapping
|
H A D | tlbflush_64.h | 19 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr); 48 void __flush_tlb_page(unsigned long context, unsigned long vaddr); 53 static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) argument 55 __flush_tlb_page(CTX_HWBITS(mm->context), vaddr); 61 void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr); 63 #define global_flush_tlb_page(mm, vaddr) \ 64 smp_flush_tlb_page(mm, vaddr)
|
H A D | viking.h | 212 static inline unsigned long viking_hwprobe(unsigned long vaddr) argument 216 vaddr &= PAGE_MASK; 220 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); 227 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); 229 vaddr &= ~SRMMU_PGDIR_MASK; 230 vaddr >>= PAGE_SHIFT; 231 return val | (vaddr << 8); 237 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); 239 vaddr &= ~SRMMU_REAL_PMD_MASK; 240 vaddr >> [all...] |
/arch/sparc/kernel/ |
H A D | iommu.c | 413 u32 vaddr, unsigned long ctx, unsigned long npages, 448 for (i = 0; i < npages; i++, vaddr += IO_PAGE_SIZE) 449 iommu_write(strbuf->strbuf_pflush, vaddr); 474 "vaddr[%08x] ctx[%lx] npages[%ld]\n", 475 vaddr, ctx, npages); 651 unsigned long vaddr, npages, entry, j; local 654 vaddr = s->dma_address & IO_PAGE_MASK; 657 iommu_range_free(iommu, vaddr, npages); 659 entry = (vaddr - iommu->page_table_map_base) 412 strbuf_flush(struct strbuf *strbuf, struct iommu *iommu, u32 vaddr, unsigned long ctx, unsigned long npages, enum dma_data_direction direction) argument
|