/arch/h8300/include/asm/ |
H A D | virtconvert.h | 13 #define phys_to_virt(vaddr) ((void *) (vaddr)) 14 #define virt_to_phys(vaddr) ((unsigned long) (vaddr))
|
/arch/frv/include/asm/ |
H A D | virtconvert.h | 24 #define phys_to_virt(vaddr) ((void *) ((unsigned long)(vaddr) + PAGE_OFFSET)) 25 #define virt_to_phys(vaddr) ((unsigned long) (vaddr) - PAGE_OFFSET) 29 #define phys_to_virt(vaddr) ((void *) (vaddr)) 30 #define virt_to_phys(vaddr) ((unsigned long) (vaddr))
|
/arch/hexagon/include/asm/ |
H A D | fixmap.h | 63 static inline unsigned long virt_to_fix(const unsigned long vaddr) argument 65 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); 66 return __virt_to_fix(vaddr); 69 #define kmap_get_fixmap_pte(vaddr) \ 70 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), \ 71 (vaddr)), (vaddr)), (vaddr))
|
/arch/m68k/include/asm/ |
H A D | bitops.h | 30 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) argument 32 char *p = (char *)vaddr + (nr ^ 31) / 8; 40 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 42 char *p = (char *)vaddr + (nr ^ 31) / 8; 49 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) argument 53 : "d" (nr ^ 31), "o" (*vaddr) 58 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) 60 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) 76 bclr_reg_clear_bit(int nr, volatile unsigned long *vaddr) argument 86 bclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 95 bfclr_mem_clear_bit(int nr, volatile unsigned long *vaddr) argument 116 bchg_reg_change_bit(int nr, volatile unsigned long *vaddr) argument 126 bchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 135 bfchg_mem_change_bit(int nr, volatile unsigned long *vaddr) argument 156 test_bit(int nr, const unsigned long *vaddr) argument 162 bset_reg_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 175 bset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 187 bfset_mem_test_and_set_bit(int nr, volatile unsigned long *vaddr) argument 212 bclr_reg_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 225 bclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 237 bfclr_mem_test_and_clear_bit(int nr, volatile unsigned long *vaddr) argument 262 bchg_reg_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 275 bchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 287 bfchg_mem_test_and_change_bit(int nr, volatile unsigned long *vaddr) argument 323 find_first_zero_bit(const unsigned long *vaddr, unsigned size) argument 349 find_next_zero_bit(const unsigned long *vaddr, int size, int offset) argument 379 find_first_bit(const unsigned long *vaddr, unsigned size) argument 404 find_next_bit(const unsigned long *vaddr, int size, int offset) argument [all...] |
/arch/ia64/xen/ |
H A D | xencomm.c | 42 xencomm_vtop(unsigned long vaddr) argument 47 if (vaddr == 0) 50 if (REGION_NUMBER(vaddr) == 5) { 60 if (vaddr >= KERNEL_START 61 && vaddr < (KERNEL_START + KERNEL_TR_PAGE_SIZE)) 62 return vaddr - kernel_virtual_offset; 65 pgd = pgd_offset_k(vaddr); 69 pud = pud_offset(pgd, vaddr); 73 pmd = pmd_offset(pud, vaddr); 77 ptep = pte_offset_kernel(pmd, vaddr); [all...] |
/arch/sparc/mm/ |
H A D | highmem.c | 35 unsigned long vaddr; local 45 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 49 __flush_cache_one(vaddr); 60 __flush_tlb_one(vaddr); 65 return (void*) vaddr; 71 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 74 if (vaddr < FIXADDR_START) { // FIXME 86 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN+idx)); 90 __flush_cache_one(vaddr); 99 pte_clear(&init_mm, vaddr, kmap_pt 117 unsigned long idx, vaddr = (unsigned long)ptr; local [all...] |
/arch/m68k/sun3/ |
H A D | dvma.c | 22 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) argument 34 // kaddr, vaddr, pte, len); 35 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { 36 sun3_put_pte(vaddr, pte); 37 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; 40 return (vaddr + (kaddr & ~PAGE_MASK)); 49 unsigned long vaddr; local 51 vaddr = dvma_btov(baddr); 53 end = vaddr + len; 55 while(vaddr < en [all...] |
/arch/sh/mm/ |
H A D | kmap.c | 18 #define kmap_get_fixmap_pte(vaddr) \ 19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr)) 25 unsigned long vaddr; local 28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); 29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); 35 unsigned long vaddr; local 45 vaddr 56 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; local 63 local_flush_tlb_one(get_asid(), vaddr); local [all...] |
/arch/arm/kernel/ |
H A D | crash_dump.c | 37 void *vaddr; local 42 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 43 if (!vaddr) 47 if (copy_to_user(buf, vaddr + offset, csize)) { 48 iounmap(vaddr); 52 memcpy(buf, vaddr + offset, csize); 55 iounmap(vaddr);
|
/arch/x86/kernel/ |
H A D | crash_dump_64.c | 29 void *vaddr; local 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); 35 if (!vaddr) 39 if (copy_to_user(buf, vaddr + offset, csize)) { 40 iounmap(vaddr); 44 memcpy(buf, vaddr + offset, csize); 47 iounmap(vaddr);
|
H A D | crash_dump_32.c | 53 void *vaddr; local 61 vaddr = kmap_atomic_pfn(pfn); 64 memcpy(buf, (vaddr + offset), csize); 65 kunmap_atomic(vaddr); 70 kunmap_atomic(vaddr); 73 copy_page(kdump_buf_page, vaddr); 74 kunmap_atomic(vaddr);
|
/arch/arm/mm/ |
H A D | highmem.c | 42 unsigned long vaddr; local 66 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 72 BUG_ON(!pte_none(get_top_pte(vaddr))); 79 set_top_pte(vaddr, mk_pte(page, kmap_prot)); 81 return (void *)vaddr; 87 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 95 __cpuc_flush_dcache_area((void *)vaddr, PAGE_SIZE); 97 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 98 set_top_pte(vaddr, __pte(0)); 103 } else if (vaddr > 113 unsigned long vaddr; local 131 unsigned long vaddr = (unsigned long)ptr; local [all...] |
H A D | cache-xsc3l2.c | 100 unsigned long vaddr; local 107 vaddr = -1; /* to force the first mapping */ 113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); 114 xsc3_l2_clean_mva(vaddr); 115 xsc3_l2_inv_mva(vaddr); 123 vaddr = l2_map_va(start, vaddr); 124 xsc3_l2_inv_mva(vaddr); 132 vaddr 144 unsigned long vaddr; local 182 unsigned long vaddr; local [all...] |
/arch/frv/mm/ |
H A D | cache-page.c | 25 void *vaddr; local 29 vaddr = kmap_atomic_primary(page, __KM_CACHE); 31 frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE); 33 kunmap_atomic_primary(vaddr, __KM_CACHE); 53 void *vaddr; local 57 vaddr = kmap_atomic_primary(page, __KM_CACHE); 59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; 62 kunmap_atomic_primary(vaddr, __KM_CACHE);
|
/arch/mips/include/asm/ |
H A D | fixmap.h | 101 static inline unsigned long virt_to_fix(const unsigned long vaddr) argument 103 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); 104 return __virt_to_fix(vaddr); 107 #define kmap_get_fixmap_pte(vaddr) \ 108 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr))
|
/arch/sh/kernel/ |
H A D | crash_dump.c | 28 void *vaddr; local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); 36 if (copy_to_user(buf, (vaddr + offset), csize)) { 37 iounmap(vaddr); 41 memcpy(buf, (vaddr + offset), csize); 43 iounmap(vaddr);
|
/arch/arm/include/asm/ |
H A D | fixmap.h | 35 static inline unsigned int virt_to_fix(const unsigned long vaddr) argument 37 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); 38 return __virt_to_fix(vaddr);
|
/arch/ia64/kernel/ |
H A D | crash_dump.c | 37 void *vaddr; local 41 vaddr = __va(pfn<<PAGE_SHIFT); 43 if (copy_to_user(buf, (vaddr + offset), csize)) { 47 memcpy(buf, (vaddr + offset), csize);
|
/arch/sparc/include/asm/ |
H A D | cache.h | 40 unsigned int vaddr, retval; local 42 vaddr = ((setnum&1) << 12) | ((tagnum&0x7f) << 5); 45 "r" (vaddr), "i" (ASI_M_TXTC_TAG)); 51 unsigned int vaddr; local 53 vaddr = ((setnum&1) << 12) | ((tagnum&0x7f) << 5); 55 "r" (entry), "r" (vaddr), "i" (ASI_M_TXTC_TAG) : 65 unsigned int value1, value2, vaddr; local 67 vaddr = ((setnum&0x1) << 12) | ((tagnum&0x7f) << 5) | 73 "r" (vaddr), "i" (ASI_M_TXTC_DATA) : 81 unsigned int value1, value2, vaddr; local [all...] |
H A D | viking.h | 212 static inline unsigned long viking_hwprobe(unsigned long vaddr) argument 216 vaddr &= PAGE_MASK; 220 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); 227 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); 229 vaddr &= ~SRMMU_PGDIR_MASK; 230 vaddr >>= PAGE_SHIFT; 231 return val | (vaddr << 8); 237 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); 239 vaddr &= ~SRMMU_REAL_PMD_MASK; 240 vaddr >> [all...] |
/arch/mips/mm/ |
H A D | pgtable-32.c | 35 unsigned long vaddr; local 54 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; 55 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); 61 vaddr = PKMAP_BASE; 62 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); 64 pgd = swapper_pg_dir + __pgd_offset(vaddr); 65 pud = pud_offset(pgd, vaddr); 66 pmd = pmd_offset(pud, vaddr); [all...] |
H A D | highmem.c | 46 unsigned long vaddr; local 56 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 61 local_flush_tlb_one((unsigned long)vaddr); 63 return (void*) vaddr; 69 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 72 if (vaddr < FIXADDR_START) { // FIXME 82 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 88 pte_clear(&init_mm, vaddr, kmap_pte-idx); 89 local_flush_tlb_one(vaddr); 103 unsigned long vaddr; local 119 unsigned long idx, vaddr = (unsigned long)ptr; local [all...] |
/arch/avr32/include/asm/ |
H A D | cacheflush.h | 20 * Invalidate any cacheline containing virtual address vaddr without 26 static inline void invalidate_dcache_line(void *vaddr) argument 30 : "r"(vaddr), "n"(CACHE_OP_DCACHE_INVALIDATE) 35 * Make sure any cacheline containing virtual address vaddr is written 38 static inline void clean_dcache_line(void *vaddr) argument 42 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN) 47 * Make sure any cacheline containing virtual address vaddr is written 50 static inline void flush_dcache_line(void *vaddr) argument 54 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN_INVAL) 60 * vaddr 62 invalidate_icache_line(void *vaddr) argument 125 copy_from_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument [all...] |
/arch/microblaze/mm/ |
H A D | highmem.c | 37 unsigned long vaddr; local 48 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); 53 local_flush_tlb_page(NULL, vaddr); 55 return (void *) vaddr; 61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 64 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { 75 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 81 pte_clear(&init_mm, vaddr, kmap_pte-idx); 82 local_flush_tlb_page(NULL, vaddr); [all...] |
/arch/powerpc/mm/ |
H A D | highmem.c | 34 unsigned long vaddr; local 44 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); 48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); 49 local_flush_tlb_page(NULL, vaddr); 51 return (void*) vaddr; 57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; local 60 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { 72 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); 78 pte_clear(&init_mm, vaddr, kmap_pte-idx); 79 local_flush_tlb_page(NULL, vaddr); [all...] |