/arch/powerpc/include/asm/ |
H A D | pgtable-ppc64-4k.h | 55 #define pgd_none(pgd) (!pgd_val(pgd)) 56 #define pgd_bad(pgd) (pgd_val(pgd) == 0) 57 #define pgd_present(pgd) (pgd_val(pgd) != 0) 59 #define pgd_page_vaddr(pgd) (pgd_val(pgd) & ~PGD_MASKED_BITS) 60 #define pgd_page(pgd) virt_to_page(pgd_page_vaddr(pgd)) [all...] |
/arch/arm/mm/ |
H A D | pgd.c | 2 * linux/arch/arm/mm/pgd.c 24 #define __pgd_free(pgd) kfree(pgd) 27 #define __pgd_free(pgd) free_pages((unsigned long)pgd, 2) 109 pgd_t *pgd; local 117 pgd = pgd_base + pgd_index(0); 118 if (pgd_none_or_clear_bad(pgd)) 121 pud = pud_offset(pgd, 0); 136 pgd_clear(pgd); [all...] |
/arch/arm/plat-samsung/include/plat/ |
H A D | sysmmu.h | 35 * #pgd: Base physical address of the 1st level page table 40 void s5p_sysmmu_enable(sysmmu_ips ips, unsigned long pgd); 54 * @pgd: The page table base address. 60 void s5p_sysmmu_set_tablebase_pgd(sysmmu_ips ips, unsigned long pgd); 89 #define s5p_sysmmu_enable(ips, pgd) do { } while (0) 91 #define s5p_sysmmu_set_tablebase_pgd(ips, pgd) do { } while (0)
|
/arch/mn10300/mm/ |
H A D | pgtable.c | 37 pgd_t *pgd; local 49 pgd = swapper_pg_dir + pgd_index(vaddr); 50 if (pgd_none(*pgd)) { 54 pud = pud_offset(pgd, vaddr); 87 * List of all pgd's needed for non-PAE so it can invalidate entries 88 * in both cached and uncached pgd's; not needed for PAE since the 103 static inline void pgd_list_add(pgd_t *pgd) argument 105 struct page *page = virt_to_page(pgd); 113 static inline void pgd_list_del(pgd_t *pgd) argument 115 struct page *next, **pprev, *page = virt_to_page(pgd); 123 pgd_ctor(void *pgd) argument 143 pgd_dtor(void *pgd) argument 157 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument [all...] |
/arch/hexagon/include/asm/ |
H A D | pgalloc.h | 36 pgd_t *pgd; local 38 pgd = (pgd_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO); 48 memcpy(pgd, swapper_pg_dir, PTRS_PER_PGD*sizeof(pgd_t *)); 52 mm->context.ptbase = __pa(pgd); 54 return pgd; 57 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 59 free_page((unsigned long) pgd); 133 pmdindex = (pgd_t *)pmd - mm->pgd; 134 ppmd = (pmd_t *)current->active_mm->pgd + pmdindex;
|
/arch/frv/mm/ |
H A D | pgalloc.c | 69 * List of all pgd's needed for non-PAE so it can invalidate entries 70 * in both cached and uncached pgd's; not needed for PAE since the 85 static inline void pgd_list_add(pgd_t *pgd) argument 87 struct page *page = virt_to_page(pgd); 95 static inline void pgd_list_del(pgd_t *pgd) argument 97 struct page *next, **pprev, *page = virt_to_page(pgd); 105 void pgd_ctor(void *pgd) argument 112 memcpy((pgd_t *) pgd + USER_PGDS_IN_LAST_PML4, 119 pgd_list_add(pgd); 121 memset(pgd, 125 pgd_dtor(void *pgd) argument 139 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument [all...] |
/arch/x86/mm/ |
H A D | pgtable.c | 73 static inline void pgd_list_add(pgd_t *pgd) argument 75 struct page *page = virt_to_page(pgd); 80 static inline void pgd_list_del(pgd_t *pgd) argument 82 struct page *page = virt_to_page(pgd); 91 static void pgd_set_mm(pgd_t *pgd, struct mm_struct *mm) argument 93 BUILD_BUG_ON(sizeof(virt_to_page(pgd)->index) < sizeof(mm)); 94 virt_to_page(pgd)->index = (pgoff_t)mm; 102 static void pgd_ctor(struct mm_struct *mm, pgd_t *pgd) argument 104 /* If the pgd points to a shared pagetable level (either the 110 clone_pgd_range(pgd 122 pgd_dtor(pgd_t *pgd) argument 220 pgd_t pgd = pgdp[i]; local 233 pgd_prepopulate_pmd(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmds[]) argument 258 pgd_t *pgd; local 296 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument [all...] |
H A D | pgtable_32.c | 29 pgd_t *pgd; local 34 pgd = swapper_pg_dir + pgd_index(vaddr); 35 if (pgd_none(*pgd)) { 39 pud = pud_offset(pgd, vaddr); 70 pgd_t *pgd; local 82 pgd = swapper_pg_dir + pgd_index(vaddr); 83 if (pgd_none(*pgd)) { 87 pud = pud_offset(pgd, vaddr);
|
/arch/parisc/include/asm/ |
H A D | pgalloc.h | 12 /* Allocate the top level pgd (page directory) 15 * allocate the first pmd adjacent to the pgd. This means that we can 16 * subtract a constant offset to get to it. The pmd and pgd sizes are 23 pgd_t *pgd = (pgd_t *)__get_free_pages(GFP_KERNEL, local 25 pgd_t *actual_pgd = pgd; 27 if (likely(pgd != NULL)) { 28 memset(pgd, 0, PAGE_SIZE<<PGD_ALLOC_ORDER); 37 + (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT)); 40 __pgd_val_set(*pgd, PxD_FLAG_ATTACHED); 46 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 58 pgd_populate(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmd) argument [all...] |
/arch/sh/mm/ |
H A D | hugetlbpage.c | 27 pgd_t *pgd; local 32 pgd = pgd_offset(mm, addr); 33 if (pgd) { 34 pud = pud_alloc(mm, pgd, addr); 47 pgd_t *pgd; local 52 pgd = pgd_offset(mm, addr); 53 if (pgd) { 54 pud = pud_offset(pgd, addr);
|
/arch/unicore32/include/asm/ |
H A D | cpu-single.h | 21 #define cpu_switch_mm(pgd, mm) cpu_do_switch_mm(virt_to_phys(pgd), mm)
|
/arch/m32r/include/asm/ |
H A D | pgtable-2level.h | 26 printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e)) 30 * setup: the pgd is never bad, and a pmd always exists (as it's folded 31 * into the pgd entry) 33 static inline int pgd_none(pgd_t pgd) { return 0; } argument 34 static inline int pgd_bad(pgd_t pgd) { return 0; } argument 35 static inline int pgd_present(pgd_t pgd) { return 1; } argument 53 #define pgd_page_vaddr(pgd) \ 54 ((unsigned long) __va(pgd_val(pgd) & PAGE_MASK)) 57 #define pgd_page(pgd) (mem_map + ((pgd_val(pgd) >> PAGE_SHIF [all...] |
H A D | pgalloc.h | 23 pgd_t *pgd = (pgd_t *)__get_free_page(GFP_KERNEL|__GFP_ZERO); local 25 return pgd; 28 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 30 free_page((unsigned long)pgd); 65 * inside the pgd, so has no extra memory associated with it. 66 * (In the PAE case we free the pmds as part of the pgd.)
|
/arch/mips/mm/ |
H A D | hugetlbpage.c | 28 pgd_t *pgd; local 32 pgd = pgd_offset(mm, addr); 33 pud = pud_alloc(mm, pgd, addr); 42 pgd_t *pgd; local 46 pgd = pgd_offset(mm, addr); 47 if (pgd_present(*pgd)) { 48 pud = pud_offset(pgd, addr);
|
/arch/ia64/xen/ |
H A D | xencomm.c | 51 pgd_t *pgd; local 65 pgd = pgd_offset_k(vaddr); 66 if (pgd_none(*pgd) || pgd_bad(*pgd)) 69 pud = pud_offset(pgd, vaddr);
|
/arch/unicore32/kernel/ |
H A D | hibernate.c | 35 static pmd_t *resume_one_md_table_init(pgd_t *pgd) argument 40 pud = pud_offset(pgd, 0); 75 pgd_t *pgd; local 81 pgd = pgd_base + pgd_idx; 84 for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) { 85 pmd = resume_one_md_table_init(pgd);
|
/arch/sh/include/asm/ |
H A D | mmu_context_32.h | 50 static inline void set_TTB(pgd_t *pgd) argument 52 __raw_writel((unsigned long)pgd, MMU_TTB);
|
H A D | mmu_context_64.h | 75 #define set_TTB(pgd) (mmu_pdtp_cache = (pgd))
|
/arch/unicore32/mm/ |
H A D | mm.h | 20 static inline pmd_t *pmd_off(pgd_t *pgd, unsigned long virt) argument 22 return pmd_offset((pud_t *)pgd, virt);
|
/arch/x86/include/asm/ |
H A D | mmzone_32.h | 17 extern void resume_map_numa_kva(pgd_t *pgd); 21 static inline void resume_map_numa_kva(pgd_t *pgd) {} argument
|
/arch/x86/power/ |
H A D | hibernate_32.c | 35 static pmd_t *resume_one_md_table_init(pgd_t *pgd) argument 45 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); 46 pud = pud_offset(pgd, 0); 50 pud = pud_offset(pgd, 0); 86 pgd_t *pgd; local 92 pgd = pgd_base + pgd_idx; 95 for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) { 96 pmd = resume_one_md_table_init(pgd);
|
/arch/um/kernel/ |
H A D | mem.c | 118 pgd_t *pgd; local 127 pgd = pgd_base + i; 129 for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) { 130 pud = pud_offset(pgd, vaddr); 163 pgd_t *pgd; local 175 pgd = swapper_pg_dir + pgd_index(vaddr); 176 pud = pud_offset(pgd, vaddr); 189 pgd_t *pgd; local 205 pgd = swapper_pg_dir + pgd_index(vaddr); 206 pud = pud_offset(pgd, vadd 273 pgd_t *pgd = (pgd_t *)__get_free_page(GFP_KERNEL); local 284 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument [all...] |
/arch/tile/mm/ |
H A D | pgtable.c | 92 pgd_t *pgd; local 97 pgd = swapper_pg_dir + pgd_index(vaddr); 98 if (pgd_none(*pgd)) { 102 pud = pud_offset(pgd, vaddr); 160 pgd_t *pgd; local 172 pgd = swapper_pg_dir + pgd_index(addr); 173 pud = pud_offset(pgd, addr); 192 /* Walk every pgd on the system and update the pmd there. */ 196 pgd = list_to_pgd(pos) + pgd_index(addr); 197 pud = pud_offset(pgd, add 229 pgd_list_add(pgd_t *pgd) argument 234 pgd_list_del(pgd_t *pgd) argument 242 pgd_ctor(pgd_t *pgd) argument 266 pgd_dtor(pgd_t *pgd) argument 277 pgd_t *pgd = kmem_cache_alloc(pgd_cache, GFP_KERNEL); local 283 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 395 pgd_t *pgd; local [all...] |
/arch/alpha/include/asm/ |
H A D | pgalloc.h | 27 pgd_populate(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmd) argument 29 pgd_set(pgd, pmd); 35 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument 37 free_page((unsigned long)pgd);
|
/arch/mips/include/asm/mach-powertv/ |
H A D | dma-coherence.h | 26 pgd_t *pgd; local 35 pgd = pgd_offset_k(virt_addr); 37 if (!pgd_none(*pgd)) { 39 pud = pud_offset(pgd, virt_addr);
|