/arch/ia64/kvm/ |
H A D | process.c | 358 * vaddr: Optional, if t == REGISTER 946 u64 oldrid, moldrid, oldpsbits, vaddr; local 948 vaddr = p->vaddr; 953 VMX(v, psbits[0]) = VMX(v, psbits[REGION_NUMBER(vaddr)]); 958 vaddr = PAGEALIGN(vaddr, p->ps); 959 thash_purge_entries_remote(v, vaddr, p->ps);
|
/arch/parisc/mm/ |
H A D | init.c | 383 static int __init parisc_text_address(unsigned long vaddr) argument 391 return core_kernel_text(vaddr) || vaddr == head_ptr; 407 unsigned long vaddr; local 432 vaddr = start_vaddr; 482 else if (parisc_text_address(vaddr) && 505 vaddr += PAGE_SIZE; 707 /* Map each physical memory range to its kernel vaddr */
|
/arch/sparc/lib/ |
H A D | copy_page.S | 47 copy_user_page: /* %o0=dest, %o1=src, %o2=vaddr */ 60 and %o2, %o3, %o0 ! vaddr D-cache alias bit 67 add %o0, %o3, %o0 ! dest TTE vaddr 70 add %o0, %o1, %o1 ! src TTE vaddr
|
/arch/x86/kernel/ |
H A D | alternative.c | 556 char *vaddr; local 573 vaddr = (char *)fix_to_virt(FIX_TEXT_POKE0); 574 memcpy(&vaddr[(unsigned long)addr & ~PAGE_MASK], opcode, len);
|
/arch/avr32/mm/ |
H A D | cache.c | 156 unsigned long vaddr, void *dst, const void *src, 155 copy_to_user_page(struct vm_area_struct *vma, struct page *page, unsigned long vaddr, void *dst, const void *src, unsigned long len) argument
|
/arch/frv/mb93090-mb00/ |
H A D | pci-dma-nommu.c | 89 void dma_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle) argument
|
/arch/m68k/include/asm/ |
H A D | dma-mapping.h | 49 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
/arch/powerpc/include/asm/ |
H A D | iommu.h | 152 void *vaddr, dma_addr_t dma_handle);
|
/arch/powerpc/kvm/ |
H A D | book3s_emulate.c | 272 ulong addr, vaddr; local 283 vaddr = addr; 288 kvmppc_set_dar(vcpu, vaddr); 289 vcpu->arch.fault_dar = vaddr;
|
/arch/sparc/include/asm/ |
H A D | leon.h | 199 unsigned long leon_swprobe(unsigned long vaddr, unsigned long *paddr);
|
/arch/tile/include/asm/ |
H A D | dma-mapping.h | 153 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, argument
|
/arch/x86/include/asm/ |
H A D | pgtable_64.h | 44 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte);
|
H A D | pgtable.h | 28 #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page)) 432 pmd_t *populate_extra_pmd(unsigned long vaddr); 433 pte_t *populate_extra_pte(unsigned long vaddr);
|
/arch/frv/kernel/ |
H A D | gdb-stub.c | 430 static inline int gdbstub_addr_probe(const void *vaddr) argument 435 asm("lrad %1,%0,#1,#0,#0" : "=r"(paddr) : "r"(vaddr)); 446 static inline unsigned long gdbstub_virt_to_pte(unsigned long vaddr) argument 454 pgd = (pgd_t *) __get_DAMLR(3) + pgd_index(vaddr); 455 pud = pud_offset(pgd, vaddr); 456 pmd = pmd_offset(pud, vaddr); 467 pte = (pte_t *)__get_DAMLR(5) + __pte_index(vaddr); 480 static inline int gdbstub_addr_map(const void *vaddr) argument 488 if (gdbstub_addr_probe(vaddr)) 491 pte = gdbstub_virt_to_pte((unsigned long) vaddr); [all...] |
/arch/score/include/asm/ |
H A D | pgtable.h | 178 #define ZERO_PAGE(vaddr) \ 180 (((unsigned long)(vaddr)) & zero_page_mask))))
|
/arch/x86/kvm/ |
H A D | paging_tmpl.h | 878 static gpa_t FNAME(gva_to_gpa)(struct kvm_vcpu *vcpu, gva_t vaddr, u32 access, argument 885 r = FNAME(walk_addr)(&walker, vcpu, vaddr, access); 889 gpa |= vaddr & ~PAGE_MASK; 897 static gpa_t FNAME(gva_to_gpa_nested)(struct kvm_vcpu *vcpu, gva_t vaddr, argument 905 r = FNAME(walk_addr_nested)(&walker, vcpu, vaddr, access); 909 gpa |= vaddr & ~PAGE_MASK;
|
/arch/arm/mm/ |
H A D | mmu.c | 403 unsigned long vaddr = __fix_to_virt(idx); local 404 pte_t *pte = pte_offset_kernel(pmd_off_k(vaddr), vaddr); 412 set_pte_at(NULL, vaddr, pte, 415 pte_clear(NULL, vaddr, pte); 416 local_flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE);
|
H A D | dma-mapping.c | 785 void *vaddr; local 794 vaddr = kmap_atomic(page); 795 op(vaddr + offset, len, dir); 796 kunmap_atomic(vaddr); 798 vaddr = kmap_high_get(page); 799 if (vaddr) { 800 op(vaddr + offset, len, dir); 805 vaddr = page_address(page) + offset; 806 op(vaddr, len, dir);
|
/arch/sparc/kernel/ |
H A D | ioport.c | 138 unsigned long vaddr = (unsigned long) virtual & PAGE_MASK; local 145 if ((res = lookup_resource(&sparc_iomap, vaddr)) == NULL) { 146 printk("free_io/iounmap: cannot free %lx\n", vaddr);
|
/arch/tile/kernel/ |
H A D | setup.c | 286 unsigned long vaddr = PAGE_OFFSET; local 325 for (j = 0; vaddr < vaddr_end; vaddr += HPAGE_SIZE, ++j) { 328 pbase_map[vaddr >> HPAGE_SHIFT] = this_pfn; 332 (void *)(vaddr & HPAGE_MASK); 339 return (void *)vaddr;
|
/arch/arc/mm/ |
H A D | tlbex.S | 234 lr r3,[ARC_REG_TLBPD0] ; MMU prepares PD0 with vaddr and asid 236 or r3, r3, r2 ; S | vaddr | {sasid|asid} 245 /* Get free TLB slot: Set = computed from vaddr, way = random */
|
/arch/frv/include/asm/ |
H A D | pgtable.h | 69 #define ZERO_PAGE(vaddr) ({ BUG(); NULL; }) 89 #define ZERO_PAGE(vaddr) virt_to_page(empty_zero_page)
|
/arch/mips/kernel/ |
H A D | traps.c | 507 unsigned long value, __user *vaddr; local 520 vaddr = (unsigned long __user *) 523 if ((unsigned long)vaddr & 3) 525 if (get_user(value, vaddr)) 546 unsigned long __user *vaddr; local 560 vaddr = (unsigned long __user *) 564 if ((unsigned long)vaddr & 3) 577 if (put_user(regs->regs[reg], vaddr))
|
/arch/mn10300/include/asm/ |
H A D | pgtable.h | 41 #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page)) 469 static inline int set_kernel_exec(unsigned long vaddr, int enable) argument
|
/arch/sparc/mm/ |
H A D | fault_64.c | 74 static void __kprobes bad_kernel_pc(struct pt_regs *regs, unsigned long vaddr) argument 80 printk(KERN_CRIT "OOPS: Fault was to vaddr[%lx]\n", vaddr);
|