Searched defs:vaddr (Results 101 - 125 of 229) sorted by relevance

12345678910

/arch/openrisc/kernel/
H A Ddma.c119 or1k_dma_free(struct device *dev, size_t size, void *vaddr, argument
122 unsigned long va = (unsigned long)vaddr;
133 free_pages_exact(vaddr, size);
/arch/parisc/include/asm/
H A Dcacheflush.h51 static inline void flush_kernel_vmap_range(void *vaddr, int size) argument
53 unsigned long start = (unsigned long)vaddr;
57 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) argument
59 unsigned long start = (unsigned long)vaddr;
60 void *cursor = vaddr;
62 for ( ; cursor < vaddr + size; cursor += PAGE_SIZE) {
92 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
99 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
101 flush_cache_page(vma, vaddr, page_to_pf
[all...]
/arch/powerpc/kernel/
H A Dio-workarounds.c28 static struct iowa_bus *iowa_pci_find(unsigned long vaddr, unsigned long paddr) argument
38 if (vaddr) {
41 if ((vaddr >= vstart) && (vaddr <= vend))
68 unsigned long vaddr, paddr; local
71 vaddr = (unsigned long)PCI_FIX_ADDR(addr);
72 if (vaddr < PHB_IO_BASE || vaddr >= PHB_IO_END)
75 ptep = find_linux_pte_or_hugepte(init_mm.pgd, vaddr,
86 bus = iowa_pci_find(vaddr, padd
104 unsigned long vaddr = (unsigned long)pci_io_base + port; local
[all...]
/arch/s390/include/asm/
H A Didals.h33 idal_is_needed(void *vaddr, unsigned int length) argument
36 return ((__pa(vaddr) + length - 1) >> 31) != 0;
46 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) argument
48 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length +
56 void *vaddr, unsigned int length)
61 paddr = __pa(vaddr);
78 set_normalized_cda(struct ccw1 * ccw, void *vaddr) argument
86 nridaws = idal_nr_words(vaddr, ccw->count);
92 idal_create_words(idal, vaddr, ccw->count);
94 vaddr
55 idal_create_words(unsigned long *idaws, void *vaddr, unsigned int length) argument
[all...]
/arch/sh/mm/
H A Dconsistent.c68 void *vaddr, dma_addr_t dma_handle,
78 iounmap(vaddr);
81 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
87 (void *)CAC_ADDR((unsigned long)vaddr) : vaddr;
67 dma_generic_free_coherent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) argument
/arch/sparc/mm/
H A Dfault_64.c74 static void __kprobes bad_kernel_pc(struct pt_regs *regs, unsigned long vaddr) argument
80 printk(KERN_CRIT "OOPS: Fault was to vaddr[%lx]\n", vaddr);
H A Dinit_32.c354 unsigned long vaddr = (unsigned long)page_address(page); local
356 if (vaddr)
357 __flush_page_to_ram(vaddr);
H A Dtlb.c69 static void tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, argument
75 vaddr &= PAGE_MASK;
77 vaddr |= 0x1UL;
87 flush_tsb_user_page(mm, vaddr);
88 global_flush_tlb_page(mm, vaddr);
95 tb->vaddrs[nr] = vaddr;
104 void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, argument
126 if ((paddr ^ vaddr) & (1 << 13))
132 tlb_batch_add_one(mm, vaddr, pte_exec(orig));
136 static void tlb_batch_pmd_scan(struct mm_struct *mm, unsigned long vaddr, argument
[all...]
/arch/tile/include/asm/
H A Dcacheflush.h65 struct page *page, unsigned long vaddr,
75 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
64 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
H A Ddma-mapping.h153 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
/arch/tile/mm/
H A Dhighmem.c20 #define kmap_get_pte(vaddr) \
21 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)),\
22 (vaddr)), (vaddr))
200 unsigned long vaddr; local
215 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx);
216 pte = kmap_get_pte(vaddr);
220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot));
222 return (void *)vaddr;
235 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local
252 kmap_atomic_unregister(pte_page(pteval), vaddr); local
281 unsigned long vaddr = (unsigned long)ptr; local
[all...]
/arch/unicore32/include/asm/
H A Ddma-mapping.h110 static inline void dma_cache_sync(struct device *dev, void *vaddr, argument
113 unsigned long start = (unsigned long)vaddr;
/arch/x86/include/asm/
H A Ddma-mapping.h68 void *vaddr, dma_addr_t dma_addr,
97 dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
161 void *vaddr, dma_addr_t bus,
168 if (dma_release_from_coherent(dev, get_order(size), vaddr))
171 debug_dma_free_coherent(dev, size, vaddr, bus);
173 ops->free(dev, size, vaddr, bus, attrs);
160 dma_free_attrs(struct device *dev, size_t size, void *vaddr, dma_addr_t bus, struct dma_attrs *attrs) argument
/arch/x86/kernel/
H A Dmachine_kexec_32.c103 unsigned long vaddr, unsigned long paddr)
107 pgd += pgd_index(vaddr);
112 pud = pud_offset(pgd, vaddr);
113 pmd = pmd_offset(pud, vaddr);
116 pte = pte_offset_kernel(pmd, vaddr);
101 machine_kexec_page_table_set_one( pgd_t *pgd, pmd_t *pmd, pte_t *pte, unsigned long vaddr, unsigned long paddr) argument
/arch/xtensa/mm/
H A Dcache.c64 unsigned long vaddr)
66 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) {
84 unsigned long vaddr, unsigned long *paddr)
86 if (PageHighMem(page) || !DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) {
88 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK));
95 void clear_user_highpage(struct page *page, unsigned long vaddr) argument
98 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr);
101 kmap_invalidate_coherent(page, vaddr);
108 unsigned long vaddr, struct vm_area_struct *vma)
111 void *dst_vaddr = coherent_kvaddr(dst, TLBTEMP_BASE_1, vaddr,
63 kmap_invalidate_coherent(struct page *page, unsigned long vaddr) argument
83 coherent_kvaddr(struct page *page, unsigned long base, unsigned long vaddr, unsigned long *paddr) argument
107 copy_user_highpage(struct page *dst, struct page *src, unsigned long vaddr, struct vm_area_struct *vma) argument
261 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument
297 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument
[all...]
/arch/alpha/boot/
H A Dbootpz.c79 unsigned long vaddr, kaddr; local
86 for (vaddr = vstart; vaddr <= vend; vaddr += PAGE_SIZE)
88 kaddr = (find_pa(vaddr) | PAGE_OFFSET);
92 srm_printk("OVERLAP: vaddr 0x%lx kaddr 0x%lx"
94 vaddr, kaddr, kstart, kend);
136 * a1 = return address, but we give the asm the vaddr of the PCB
/arch/alpha/kernel/
H A Dcore_irongate.c320 unsigned long vaddr;
386 for(baddr = addr, vaddr = (unsigned long)area->addr;
388 baddr += PAGE_SIZE, vaddr += PAGE_SIZE)
393 if (__alpha_remap_area_pages(vaddr,
403 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK);
406 addr, size, vaddr);
408 return (void __iomem *)vaddr;
319 unsigned long vaddr; local
H A Dcore_tsunami.c202 tsunami_probe_read(volatile unsigned long *vaddr)
211 dont_care = *vaddr;
224 tsunami_probe_write(volatile unsigned long *vaddr)
229 true_contents = *vaddr;
230 *vaddr = 0;
237 (unsigned long)vaddr);
240 *vaddr = true_contents;
200 tsunami_probe_read(volatile unsigned long *vaddr) argument
222 tsunami_probe_write(volatile unsigned long *vaddr) argument
/arch/arm/kernel/
H A Duprobes.c33 unsigned long vaddr)
35 return uprobe_write_opcode(mm, vaddr,
116 void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, argument
120 void *dst = xol_page_kaddr + (vaddr & ~PAGE_MASK);
128 flush_uprobe_xol_access(page, vaddr, dst, len);
157 regs->ARM_pc = utask->vaddr + 4;
178 instruction_pointer_set(regs, utask->vaddr);
32 set_swbp(struct arch_uprobe *auprobe, struct mm_struct *mm, unsigned long vaddr) argument
/arch/arm/mm/
H A Dcache-feroceon-l2.c54 void *vaddr = kmap_atomic_pfn(paddr >> PAGE_SHIFT); local
55 return (unsigned long)vaddr + (paddr & ~PAGE_MASK);
61 static inline void l2_put_va(unsigned long vaddr) argument
64 kunmap_atomic((void *)vaddr);
H A Dflush.c26 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) argument
28 unsigned long to = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT);
40 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) argument
42 unsigned long va = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT);
43 unsigned long offset = vaddr & (PAGE_SIZE - 1);
103 #define flush_pfn_alias(pfn,vaddr) do { } while (0)
104 #define flush_icache_alias(pfn,vaddr,len) do { } while (0)
H A Dioremap.c68 struct static_vm *find_static_vm_vaddr(void *vaddr) argument
77 if (vm->addr > vaddr)
80 if (vm->addr <= vaddr && vm->addr + vm->size > vaddr)
91 void *vaddr; local
95 vaddr = vm->addr;
100 if (vm->addr > vaddr)
/arch/m32r/mm/
H A Dfault.c346 void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, argument
361 vaddr = (vaddr & PAGE_MASK) | get_asid();
368 if (*entry1++ == vaddr) {
376 if (*entry2++ == vaddr) {
403 : "r" (vaddr), "r" (pte_data), "i" (MMU_REG_BASE),
447 *entry1++ = vaddr; /* Set TLB tag */
/arch/m68k/include/asm/
H A Dbitops.h31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) argument
33 char *p = (char *)vaddr + (nr ^ 31) / 8;
41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument
43 char *p = (char *)vaddr + (nr ^ 31) / 8;
50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument
54 : "d" (nr ^ 31), "o" (*vaddr)
59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr)
61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr)
71 bclr_reg_clear_bit(int nr, volatile unsigned long *vaddr) argument
81 bclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument
90 bfclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument
111 bchg_reg_change_bit(int nr, volatile unsigned long *vaddr) argument
121 bchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument
130 bfchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument
151 test_bit(int nr, const unsigned long *vaddr) argument
157 bset_reg_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument
170 bset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument
182 bfset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument
207 bclr_reg_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument
220 bclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument
232 bfclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument
257 bchg_reg_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument
270 bchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument
282 bfchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument
318 find_first_zero_bit(const unsigned long *vaddr, unsigned size) argument
344 find_next_zero_bit(const unsigned long *vaddr, int size, int offset) argument
374 find_first_bit(const unsigned long *vaddr, unsigned size) argument
399 find_next_bit(const unsigned long *vaddr, int size, int offset) argument
[all...]
H A Dcacheflush_mm.h155 extern void cache_push_v(unsigned long vaddr, int len);
222 static inline void __flush_page_to_ram(void *vaddr) argument
226 addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1);
239 : : "a" (__pa(vaddr)));
261 struct page *page, unsigned long vaddr,
264 flush_cache_page(vma, vaddr, page_to_pfn(page));
266 flush_icache_user_range(vma, page, vaddr, len);
269 struct page *page, unsigned long vaddr,
272 flush_cache_page(vma, vaddr, page_to_pfn(page));
260 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument
268 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, void *src, int len) argument

Completed in 1105 milliseconds

12345678910