/arch/hexagon/include/asm/ |
H A D | fixmap.h | 31 #define kmap_get_fixmap_pte(vaddr) \ 32 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), \ 33 (vaddr)), (vaddr)), (vaddr))
|
/arch/frv/include/asm/ |
H A D | virtconvert.h | 24 #define phys_to_virt(vaddr) ((void *) ((unsigned long)(vaddr) + PAGE_OFFSET)) 25 #define virt_to_phys(vaddr) ((unsigned long) (vaddr) - PAGE_OFFSET) 29 #define phys_to_virt(vaddr) ((void *) (vaddr)) 30 #define virt_to_phys(vaddr) ((unsigned long) (vaddr))
|
/arch/m68k/include/asm/ |
H A D | bitops.h | 31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) argument 33 char *p = (char *)vaddr + (nr ^ 31) / 8; 41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 43 char *p = (char *)vaddr + (nr ^ 31) / 8; 50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 54 : "d" (nr ^ 31), "o" (*vaddr) 59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) 61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) 71 bclr_reg_clear_bit(int nr, volatile unsigned long *vaddr) argument 81 bclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 90 bfclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 111 bchg_reg_change_bit(int nr, volatile unsigned long *vaddr) argument 121 bchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 130 bfchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 151 test_bit(int nr, const unsigned long *vaddr) argument 157 bset_reg_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 170 bset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 182 bfset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 207 bclr_reg_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 220 bclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 232 bfclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 257 bchg_reg_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 270 bchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 282 bfchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 318 find_first_zero_bit(const unsigned long *vaddr, unsigned size) argument 344 find_next_zero_bit(const unsigned long *vaddr, int size, int offset) argument 374 find_first_bit(const unsigned long *vaddr, unsigned size) argument 399 find_next_bit(const unsigned long *vaddr, int size, int offset) argument [all...] |
/arch/sh/mm/ |
H A D | kmap.c | 18 #define kmap_get_fixmap_pte(vaddr) \ 19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr)) 25 unsigned long vaddr; local 28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); 29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); 35 unsigned long vaddr; local 45 vaddr 56 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; local 63 local_flush_tlb_one(get_asid(), vaddr); local [all...] |
/arch/arm/kernel/ |
H A D | crash_dump.c | 37 void *vaddr; local 42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); 43 if (!vaddr) 47 if (copy_to_user(buf, vaddr + offset, csize)) { 48 iounmap(vaddr); 52 memcpy(buf, vaddr + offset, csize); 55 iounmap(vaddr);
|
/arch/x86/kernel/ |
H A D | crash_dump_64.c | 29 void *vaddr; local 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); 35 if (!vaddr) 39 if (copy_to_user(buf, vaddr + offset, csize)) { 40 iounmap(vaddr); 44 memcpy(buf, vaddr + offset, csize); 47 iounmap(vaddr);
|
H A D | crash_dump_32.c | 53 void *vaddr; local 61 vaddr = kmap_atomic_pfn(pfn); 64 memcpy(buf, (vaddr + offset), csize); 65 kunmap_atomic(vaddr); 70 kunmap_atomic(vaddr); 73 copy_page(kdump_buf_page, vaddr); 74 kunmap_atomic(vaddr);
|
/arch/arm/mm/ |
H A D | cache-xsc3l2.c | 100 unsigned long vaddr; local 107 vaddr = -1; /* to force the first mapping */ 113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); 114 xsc3_l2_clean_mva(vaddr); 115 xsc3_l2_inv_mva(vaddr); 123 vaddr = l2_map_va(start, vaddr); 124 xsc3_l2_inv_mva(vaddr); 132 vaddr 144 unsigned long vaddr; local 182 unsigned long vaddr; local [all...] |
H A D | highmem.c | 23 unsigned long vaddr = __fix_to_virt(idx); local 24 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); 27 local_flush_tlb_kernel_page(vaddr); 30 static inline pte_t get_fixmap_pte(unsigned long vaddr) argument 32 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); 58 unsigned long vaddr; local 82 vaddr = __fix_to_virt(idx); 88 BUG_ON(!pte_none(get_fixmap_pte(vaddr))); 103 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 129 unsigned long vaddr; local 150 unsigned long vaddr = (unsigned long)ptr; local [all...] |
/arch/frv/mm/ |
H A D | cache-page.c | 25 void *vaddr; local 29 vaddr = kmap_atomic_primary(page); 31 frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE); 33 kunmap_atomic_primary(vaddr); 53 void *vaddr; local 57 vaddr = kmap_atomic_primary(page); 59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; 62 kunmap_atomic_primary(vaddr);
|
/arch/xtensa/include/asm/ |
H A D | fixmap.h | 66 static inline unsigned long virt_to_fix(const unsigned long vaddr) argument 68 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); 69 return __virt_to_fix(vaddr); 74 #define kmap_get_fixmap_pte(vaddr) \ 76 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \ 77 (vaddr) \
|
/arch/sh/kernel/ |
H A D | crash_dump.c | 28 void *vaddr; local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 36 if (copy_to_user(buf, (vaddr + offset), csize)) { 37 iounmap(vaddr); 41 memcpy(buf, (vaddr + offset), csize); 43 iounmap(vaddr);
|
/arch/ia64/kernel/ |
H A D | crash_dump.c | 37 void *vaddr; local 41 vaddr = __va(pfn<<PAGE_SHIFT); 43 if (copy_to_user(buf, (vaddr + offset), csize)) { 47 memcpy(buf, (vaddr + offset), csize);
|
/arch/metag/include/asm/ |
H A D | fixmap.h | 56 #define kmap_get_fixmap_pte(vaddr) \ 58 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \ 59 (vaddr) \
|
/arch/mips/include/asm/ |
H A D | fixmap.h | 72 #define kmap_get_fixmap_pte(vaddr) \ 73 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr))
|
/arch/m68k/sun3/ |
H A D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) argument 35 // kaddr, vaddr, pte, len); 36 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { 37 sun3_put_pte(vaddr, pte); 38 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; 41 return (vaddr + (kaddr & ~PAGE_MASK)); 50 unsigned long vaddr; local 52 vaddr = dvma_btov(baddr); 54 end = vaddr + len; 56 while(vaddr < en [all...] |
/arch/mips/mm/ |
H A D | pgtable-32.c | 35 unsigned long vaddr; local 54 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; 55 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); 61 vaddr = PKMAP_BASE; 62 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); 64 pgd = swapper_pg_dir + __pgd_offset(vaddr); 65 pud = pud_offset(pgd, vaddr); 66 pmd = pmd_offset(pud, vaddr); [all...] |
H A D | highmem.c | 47 unsigned long vaddr; local 57 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 62 local_flush_tlb_one((unsigned long)vaddr); 64 return (void*) vaddr; 70 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 73 if (vaddr < FIXADDR_START) { // FIXME 83 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 89 pte_clear(&init_mm, vaddr, kmap_pte-idx); 90 local_flush_tlb_one(vaddr); 104 unsigned long vaddr; local 120 unsigned long idx, vaddr = (unsigned long)ptr; local [all...] |
/arch/sparc/mm/ |
H A D | highmem.c | 53 unsigned long vaddr; local 63 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 67 __flush_cache_one(vaddr); 78 __flush_tlb_one(vaddr); 83 return (void*) vaddr; 89 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 92 if (vaddr < FIXADDR_START) { // FIXME 104 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN+idx)); 108 __flush_cache_one(vaddr); 117 pte_clear(&init_mm, vaddr, kmap_pt [all...] |
/arch/sparc/include/asm/ |
H A D | tlbflush_64.h | 19 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr); 48 void __flush_tlb_page(unsigned long context, unsigned long vaddr); 53 static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) argument 55 __flush_tlb_page(CTX_HWBITS(mm->context), vaddr); 61 void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr); 63 #define global_flush_tlb_page(mm, vaddr) \ 64 smp_flush_tlb_page(mm, vaddr)
|
H A D | viking.h | 212 static inline unsigned long viking_hwprobe(unsigned long vaddr) argument 216 vaddr &= PAGE_MASK; 220 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); 227 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); 229 vaddr &= ~SRMMU_PGDIR_MASK; 230 vaddr >>= PAGE_SHIFT; 231 return val | (vaddr << 8); 237 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); 239 vaddr &= ~SRMMU_REAL_PMD_MASK; 240 vaddr >> [all...] |
/arch/avr32/include/asm/ |
H A D | cacheflush.h | 20 * Invalidate any cacheline containing virtual address vaddr without 26 static inline void invalidate_dcache_line(void *vaddr) argument 30 : "r"(vaddr), "n"(CACHE_OP_DCACHE_INVALIDATE) 35 * Make sure any cacheline containing virtual address vaddr is written 38 static inline void clean_dcache_line(void *vaddr) argument 42 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN) 47 * Make sure any cacheline containing virtual address vaddr is written 50 static inline void flush_dcache_line(void *vaddr) argument 54 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN_INVAL) 60 * vaddr 62 invalidate_icache_line(void *vaddr) argument 125 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument [all...] |
/arch/microblaze/mm/ |
H A D | highmem.c | 37 unsigned long vaddr; local 48 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); 53 local_flush_tlb_page(NULL, vaddr); 55 return (void *) vaddr; 61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 64 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { 75 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 81 pte_clear(&init_mm, vaddr, kmap_pte-idx); 82 local_flush_tlb_page(NULL, vaddr); [all...] |
/arch/powerpc/mm/ |
H A D | highmem.c | 34 unsigned long vaddr; local 44 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); 49 local_flush_tlb_page(NULL, vaddr); 51 return (void*) vaddr; 57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 60 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { 72 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 78 pte_clear(&init_mm, vaddr, kmap_pte-idx); 79 local_flush_tlb_page(NULL, vaddr); [all...] |
/arch/mips/kernel/ |
H A D | crash_dump.c | 29 void *vaddr; local 34 vaddr = kmap_atomic_pfn(pfn); 37 memcpy(buf, (vaddr + offset), csize); 38 kunmap_atomic(vaddr); 45 copy_page(kdump_buf_page, vaddr); 46 kunmap_atomic(vaddr);
|