Lines Matching defs:address
99 * This performs some platform-dependent address space initialization.
133 /* map NaT-page at address zero to speed up speculative dereferencing of NULL: */
212 put_kernel_page (struct page *page, unsigned long address, pgprot_t pgprot)
223 pgd = pgd_offset_k(address); /* note: this is NOT pgd_offset()! */
226 pud = pud_alloc(&init_mm, pgd, address);
229 pmd = pmd_alloc(&init_mm, pud, address);
232 pte = pte_alloc_kernel(pmd, address);
325 * address space. The IA-64 architecture guarantees that at least 50 bits of
326 * virtual address space are implemented but if we pick a large enough page size
327 * (e.g., 64KB), the mapped address space is big enough that it will overlap with
331 * address space to not permit mappings that would overlap with the VMLPT.
337 * The virtual page table has to cover the entire implemented address space within
340 * non-speculative accesses to the virtual page table, so the address range of the
359 " to cover mapped address space.\n"
367 * Set the (virtually mapped linear) page table address. Bit
439 unsigned long address, start_page, end_page;
454 for (address = start_page; address < end_page; address += PAGE_SIZE) {
455 pgd = pgd_offset_k(address);
458 pud = pud_offset(pgd, address);
462 pmd = pmd_offset(pud, address);
466 pte = pte_offset_kernel(pmd, address);