/arch/arm64/mm/ |
H A D | copypage.c | 32 void __cpu_clear_user_page(void *kaddr, unsigned long vaddr) argument 34 clear_page(kaddr);
|
H A D | flush.c | 38 unsigned long uaddr, void *kaddr, 42 unsigned long addr = (unsigned long)kaddr; 44 __flush_dcache_area(kaddr, len); 37 flush_ptrace_access(struct vm_area_struct *vma, struct page *page, unsigned long uaddr, void *kaddr, unsigned long len) argument
|
/arch/sh/mm/ |
H A D | uncached.c | 23 int virt_addr_uncached(unsigned long kaddr) argument 25 return (kaddr >= uncached_start) && (kaddr < uncached_end);
|
/arch/arm/mm/ |
H A D | copypage-fa.c | 61 void *ptr, *kaddr = kmap_atomic(page); local 78 : "0" (kaddr), "I" (PAGE_SIZE / 32) 80 kunmap_atomic(kaddr);
|
H A D | copypage-feroceon.c | 85 void *ptr, *kaddr = kmap_atomic(page); local 103 : "0" (kaddr), "I" (PAGE_SIZE / 32) 105 kunmap_atomic(kaddr);
|
H A D | copypage-v4mc.c | 91 void *ptr, *kaddr = kmap_atomic(page); local 107 : "0" (kaddr), "I" (PAGE_SIZE / 64) 109 kunmap_atomic(kaddr);
|
H A D | copypage-v4wb.c | 70 void *ptr, *kaddr = kmap_atomic(page); local 87 : "0" (kaddr), "I" (PAGE_SIZE / 64) 89 kunmap_atomic(kaddr);
|
H A D | copypage-v4wt.c | 65 void *ptr, *kaddr = kmap_atomic(page); local 80 : "0" (kaddr), "I" (PAGE_SIZE / 64) 82 kunmap_atomic(kaddr);
|
H A D | copypage-v6.c | 51 void *kaddr = kmap_atomic(page); local 52 clear_page(kaddr); 53 kunmap_atomic(kaddr);
|
H A D | copypage-xsc3.c | 93 void *ptr, *kaddr = kmap_atomic(page); local 106 : "0" (kaddr), "I" (PAGE_SIZE / 32) 108 kunmap_atomic(kaddr);
|
H A D | copypage-xscale.c | 112 void *ptr, *kaddr = kmap_atomic(page); local 127 : "0" (kaddr), "I" (PAGE_SIZE / 32) 129 kunmap_atomic(kaddr);
|
H A D | flush.c | 116 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, argument 121 unsigned long addr = (unsigned long)kaddr; 135 unsigned long addr = (unsigned long)kaddr; 148 unsigned long uaddr, void *kaddr, unsigned long len) 155 __flush_ptrace_access(page, uaddr, kaddr, len, flags); 159 void *kaddr, unsigned long len) 163 __flush_ptrace_access(page, uaddr, kaddr, len, flags); 147 flush_ptrace_access(struct vm_area_struct *vma, struct page *page, unsigned long uaddr, void *kaddr, unsigned long len) argument 158 flush_uprobe_xol_access(struct page *page, unsigned long uaddr, void *kaddr, unsigned long len) argument
|
/arch/m68k/sun3/ |
H A D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) argument 29 j = *(volatile unsigned long *)kaddr; 30 *(volatile unsigned long *)kaddr = j; 32 ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); 35 // kaddr, vaddr, pte, len); 41 return (vaddr + (kaddr & ~PAGE_MASK)); 45 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, argument 57 dvma_page(kaddr, vaddr); 58 kaddr += PAGE_SIZE;
|
H A D | sun3dvma.c | 280 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) argument 289 if(!kaddr || !len) { 290 // printk("error: kaddr %lx len %x\n", kaddr, len); 297 len, kaddr); 299 off = kaddr & ~DVMA_PAGE_MASK; 300 kaddr &= PAGE_MASK; 312 if(!dvma_map_iommu(kaddr, baddr, len)) 315 printk("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, badd 339 unsigned long kaddr; local [all...] |
/arch/unicore32/include/asm/ |
H A D | tlbflush.h | 58 * flush_kern_tlb_page(kaddr) 63 * - kaddr - Kernel virtual memory address 106 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) argument 111 : : "r" (kaddr & PAGE_MASK) : "cc"); 114 : : "r" (kaddr & PAGE_MASK) : "cc"); 118 : : "r" (kaddr & PAGE_MASK) : "cc");
|
/arch/unicore32/mm/ |
H A D | flush.c | 36 unsigned long uaddr, void *kaddr, unsigned long len) 40 unsigned long addr = (unsigned long)kaddr; 35 flush_ptrace_access(struct vm_area_struct *vma, struct page *page, unsigned long uaddr, void *kaddr, unsigned long len) argument
|
/arch/arm/kernel/ |
H A D | smp_tlb.c | 169 void flush_tlb_kernel_page(unsigned long kaddr) argument 173 ta.ta_start = kaddr; 176 __flush_tlb_kernel_page(kaddr);
|
/arch/m68k/sun3x/ |
H A D | dvma.c | 82 inline int dvma_map_cpu(unsigned long kaddr, argument 89 kaddr &= PAGE_MASK; 96 kaddr, vaddr); 131 __pa(kaddr), vaddr); 133 set_pte(pte, pfn_pte(virt_to_pfn(kaddr), 136 kaddr += PAGE_SIZE; 151 inline int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, argument 165 // printk("mapping pa %lx to ba %lx\n", __pa(kaddr), index << DVMA_PAGE_SHIFT); 167 dvma_entry_set(index, __pa(kaddr)); 172 kaddr [all...] |
/arch/mips/mm/ |
H A D | mmap.c | 194 int __virt_addr_valid(const volatile void *kaddr) argument 196 return pfn_valid(PFN_DOWN(virt_to_phys(kaddr)));
|
H A D | c-r3k.c | 239 unsigned long kaddr = KSEG0ADDR(pfn << PAGE_SHIFT); local 263 r3k_flush_dcache_range(kaddr, kaddr + PAGE_SIZE); 265 r3k_flush_icache_range(kaddr, kaddr + PAGE_SIZE);
|
H A D | cache.c | 110 void *kaddr; local 112 kaddr = kmap_coherent(page, vmaddr); 113 flush_data_cache_page((unsigned long)kaddr);
|
/arch/score/mm/ |
H A D | cache.c | 215 unsigned long kaddr = 0xa0000000 | (pfn << PAGE_SHIFT); local 217 flush_dcache_range(kaddr, kaddr + PAGE_SIZE); 220 flush_icache_range(kaddr, kaddr + PAGE_SIZE);
|
/arch/x86/lib/ |
H A D | insn.c | 31 ((insn)->next_byte + sizeof(t) + n - (insn)->kaddr <= MAX_INSN_SIZE) 50 * @kaddr: address (in kernel memory) of instruction (or copy thereof) 53 void insn_init(struct insn *insn, const void *kaddr, int x86_64) argument 56 insn->kaddr = kaddr; 57 insn->next_byte = kaddr; 579 - (unsigned long)insn->kaddr);
|
/arch/alpha/boot/ |
H A D | bootpz.c | 79 unsigned long vaddr, kaddr; local 88 kaddr = (find_pa(vaddr) | PAGE_OFFSET); 89 if (kaddr >= kstart && kaddr <= kend) 92 srm_printk("OVERLAP: vaddr 0x%lx kaddr 0x%lx" 94 vaddr, kaddr, kstart, kend);
|
/arch/metag/kernel/ |
H A D | dma.c | 214 void *kaddr = page_address(page); local 215 memset(kaddr, 0, size); 216 flush_dcache_region(kaddr, size);
|