/arch/openrisc/kernel/ |
H A D | dma.c | 119 or1k_dma_free(struct device *dev, size_t size, void *vaddr, argument 122 unsigned long va = (unsigned long)vaddr; 133 free_pages_exact(vaddr, size);
|
/arch/parisc/include/asm/ |
H A D | cacheflush.h | 51 static inline void flush_kernel_vmap_range(void *vaddr, int size) argument 53 unsigned long start = (unsigned long)vaddr; 57 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) argument 59 unsigned long start = (unsigned long)vaddr; 60 void *cursor = vaddr; 62 for ( ; cursor < vaddr + size; cursor += PAGE_SIZE) { 92 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 99 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 101 flush_cache_page(vma, vaddr, page_to_pf [all...] |
/arch/powerpc/kernel/ |
H A D | io-workarounds.c | 28 static struct iowa_bus *iowa_pci_find(unsigned long vaddr, unsigned long paddr) argument 38 if (vaddr) { 41 if ((vaddr >= vstart) && (vaddr <= vend)) 68 unsigned long vaddr, paddr; local 71 vaddr = (unsigned long)PCI_FIX_ADDR(addr); 72 if (vaddr < PHB_IO_BASE || vaddr >= PHB_IO_END) 75 ptep = find_linux_pte_or_hugepte(init_mm.pgd, vaddr, 86 bus = iowa_pci_find(vaddr, padd 104 unsigned long vaddr = (unsigned long)pci_io_base + port; local [all...] |
/arch/s390/include/asm/ |
H A D | idals.h | 33 idal_is_needed(void *vaddr, unsigned int length) argument 36 return ((__pa(vaddr) + length - 1) >> 31) != 0; 46 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) argument 48 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + 56 void *vaddr, unsigned int length) 61 paddr = __pa(vaddr); 78 set_normalized_cda(struct ccw1 * ccw, void *vaddr) argument 86 nridaws = idal_nr_words(vaddr, ccw->count); 92 idal_create_words(idal, vaddr, ccw->count); 94 vaddr 55 idal_create_words(unsigned long *idaws, void *vaddr, unsigned int length) argument [all...] |
/arch/sh/mm/ |
H A D | consistent.c | 68 void *vaddr, dma_addr_t dma_handle, 78 iounmap(vaddr); 81 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 87 (void *)CAC_ADDR((unsigned long)vaddr) : vaddr; 67 dma_generic_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
|
/arch/sparc/mm/ |
H A D | fault_64.c | 74 static void __kprobes bad_kernel_pc(struct pt_regs *regs, unsigned long vaddr) argument 80 printk(KERN_CRIT "OOPS: Fault was to vaddr[%lx]\n", vaddr);
|
H A D | init_32.c | 354 unsigned long vaddr = (unsigned long)page_address(page); local 356 if (vaddr) 357 __flush_page_to_ram(vaddr);
|
H A D | tlb.c | 69 static void tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, argument 75 vaddr &= PAGE_MASK; 77 vaddr |= 0x1UL; 87 flush_tsb_user_page(mm, vaddr); 88 global_flush_tlb_page(mm, vaddr); 95 tb->vaddrs[nr] = vaddr; 104 void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, argument 126 if ((paddr ^ vaddr) & (1 << 13)) 132 tlb_batch_add_one(mm, vaddr, pte_exec(orig)); 136 static void tlb_batch_pmd_scan(struct mm_struct *mm, unsigned long vaddr, argument [all...] |
/arch/tile/include/asm/ |
H A D | cacheflush.h | 65 struct page *page, unsigned long vaddr, 75 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 64 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
|
H A D | dma-mapping.h | 153 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
/arch/tile/mm/ |
H A D | highmem.c | 20 #define kmap_get_pte(vaddr) \ 21 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)),\ 22 (vaddr)), (vaddr)) 200 unsigned long vaddr; local 215 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 216 pte = kmap_get_pte(vaddr); 220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot)); 222 return (void *)vaddr; 235 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 252 kmap_atomic_unregister(pte_page(pteval), vaddr); local 281 unsigned long vaddr = (unsigned long)ptr; local [all...] |
/arch/unicore32/include/asm/ |
H A D | dma-mapping.h | 110 static inline void dma_cache_sync(struct device *dev, void *vaddr, argument 113 unsigned long start = (unsigned long)vaddr;
|
/arch/x86/include/asm/ |
H A D | dma-mapping.h | 68 void *vaddr, dma_addr_t dma_addr, 97 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument 161 void *vaddr, dma_addr_t bus, 168 if (dma_release_from_coherent(dev, get_order(size), vaddr)) 171 debug_dma_free_coherent(dev, size, vaddr, bus); 173 ops->free(dev, size, vaddr, bus, attrs); 160 dma_free_attrs(struct device *dev, size_t size, void *vaddr, dma_addr_t bus, struct dma_attrs *attrs) argument
|
/arch/x86/kernel/ |
H A D | machine_kexec_32.c | 103 unsigned long vaddr, unsigned long paddr) 107 pgd += pgd_index(vaddr); 112 pud = pud_offset(pgd, vaddr); 113 pmd = pmd_offset(pud, vaddr); 116 pte = pte_offset_kernel(pmd, vaddr); 101 machine_kexec_page_table_set_one( pgd_t *pgd, pmd_t *pmd, pte_t *pte, unsigned long vaddr, unsigned long paddr) argument
|
/arch/xtensa/mm/ |
H A D | cache.c | 64 unsigned long vaddr) 66 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { 84 unsigned long vaddr, unsigned long *paddr) 86 if (PageHighMem(page) || !DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { 88 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK)); 95 void clear_user_highpage(struct page *page, unsigned long vaddr) argument 98 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr); 101 kmap_invalidate_coherent(page, vaddr); 108 unsigned long vaddr, struct vm_area_struct *vma) 111 void *dst_vaddr = coherent_kvaddr(dst, TLBTEMP_BASE_1, vaddr, 63 kmap_invalidate_coherent(struct page *page, unsigned long vaddr) argument 83 coherent_kvaddr(struct page *page, unsigned long base, unsigned long vaddr, unsigned long *paddr) argument 107 copy_user_highpage(struct page *dst, struct page *src, unsigned long vaddr, struct vm_area_struct *vma) argument 261 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument 297 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument [all...] |
/arch/alpha/boot/ |
H A D | bootpz.c | 79 unsigned long vaddr, kaddr; local 86 for (vaddr = vstart; vaddr <= vend; vaddr += PAGE_SIZE) 88 kaddr = (find_pa(vaddr) | PAGE_OFFSET); 92 srm_printk("OVERLAP: vaddr 0x%lx kaddr 0x%lx" 94 vaddr, kaddr, kstart, kend); 136 * a1 = return address, but we give the asm the vaddr of the PCB
|
/arch/alpha/kernel/ |
H A D | core_irongate.c | 320 unsigned long vaddr; 386 for(baddr = addr, vaddr = (unsigned long)area->addr; 388 baddr += PAGE_SIZE, vaddr += PAGE_SIZE) 393 if (__alpha_remap_area_pages(vaddr, 403 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); 406 addr, size, vaddr); 408 return (void __iomem *)vaddr; 319 unsigned long vaddr; local
|
H A D | core_tsunami.c | 202 tsunami_probe_read(volatile unsigned long *vaddr) 211 dont_care = *vaddr; 224 tsunami_probe_write(volatile unsigned long *vaddr) 229 true_contents = *vaddr; 230 *vaddr = 0; 237 (unsigned long)vaddr); 240 *vaddr = true_contents; 200 tsunami_probe_read(volatile unsigned long *vaddr) argument 222 tsunami_probe_write(volatile unsigned long *vaddr) argument
|
/arch/arm/kernel/ |
H A D | uprobes.c | 33 unsigned long vaddr) 35 return uprobe_write_opcode(mm, vaddr, 116 void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, argument 120 void *dst = xol_page_kaddr + (vaddr & ~PAGE_MASK); 128 flush_uprobe_xol_access(page, vaddr, dst, len); 157 regs->ARM_pc = utask->vaddr + 4; 178 instruction_pointer_set(regs, utask->vaddr); 32 set_swbp(struct arch_uprobe *auprobe, struct mm_struct *mm, unsigned long vaddr) argument
|
/arch/arm/mm/ |
H A D | cache-feroceon-l2.c | 54 void *vaddr = kmap_atomic_pfn(paddr >> PAGE_SHIFT); local 55 return (unsigned long)vaddr + (paddr & ~PAGE_MASK); 61 static inline void l2_put_va(unsigned long vaddr) argument 64 kunmap_atomic((void *)vaddr);
|
H A D | flush.c | 26 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) argument 28 unsigned long to = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); 40 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) argument 42 unsigned long va = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); 43 unsigned long offset = vaddr & (PAGE_SIZE - 1); 103 #define flush_pfn_alias(pfn,vaddr) do { } while (0) 104 #define flush_icache_alias(pfn,vaddr,len) do { } while (0)
|
H A D | ioremap.c | 68 struct static_vm *find_static_vm_vaddr(void *vaddr) argument 77 if (vm->addr > vaddr) 80 if (vm->addr <= vaddr && vm->addr + vm->size > vaddr) 91 void *vaddr; local 95 vaddr = vm->addr; 100 if (vm->addr > vaddr)
|
/arch/m32r/mm/ |
H A D | fault.c | 346 void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, argument 361 vaddr = (vaddr & PAGE_MASK) | get_asid(); 368 if (*entry1++ == vaddr) { 376 if (*entry2++ == vaddr) { 403 : "r" (vaddr), "r" (pte_data), "i" (MMU_REG_BASE), 447 *entry1++ = vaddr; /* Set TLB tag */
|
/arch/m68k/include/asm/ |
H A D | bitops.h | 31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) argument 33 char *p = (char *)vaddr + (nr ^ 31) / 8; 41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 43 char *p = (char *)vaddr + (nr ^ 31) / 8; 50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 54 : "d" (nr ^ 31), "o" (*vaddr) 59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) 61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) 71 bclr_reg_clear_bit(int nr, volatile unsigned long *vaddr) argument 81 bclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 90 bfclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 111 bchg_reg_change_bit(int nr, volatile unsigned long *vaddr) argument 121 bchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 130 bfchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 151 test_bit(int nr, const unsigned long *vaddr) argument 157 bset_reg_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 170 bset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 182 bfset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 207 bclr_reg_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 220 bclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 232 bfclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 257 bchg_reg_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 270 bchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 282 bfchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 318 find_first_zero_bit(const unsigned long *vaddr, unsigned size) argument 344 find_next_zero_bit(const unsigned long *vaddr, int size, int offset) argument 374 find_first_bit(const unsigned long *vaddr, unsigned size) argument 399 find_next_bit(const unsigned long *vaddr, int size, int offset) argument [all...] |
H A D | cacheflush_mm.h | 155 extern void cache_push_v(unsigned long vaddr, int len); 222 static inline void __flush_page_to_ram(void *vaddr) argument 226 addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1); 239 : : "a" (__pa(vaddr))); 261 struct page *page, unsigned long vaddr, 264 flush_cache_page(vma, vaddr, page_to_pfn(page)); 266 flush_icache_user_range(vma, page, vaddr, len); 269 struct page *page, unsigned long vaddr, 272 flush_cache_page(vma, vaddr, page_to_pfn(page)); 260 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument 268 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
|