Searched refs:pgd (Results 1 - 25 of 270) sorted by relevance

1234567891011

/arch/powerpc/include/asm/
H A Dpgtable-ppc64-4k.h55 #define pgd_none(pgd) (!pgd_val(pgd))
56 #define pgd_bad(pgd) (pgd_val(pgd) == 0)
57 #define pgd_present(pgd) (pgd_val(pgd) != 0)
59 #define pgd_page_vaddr(pgd) (pgd_val(pgd) & ~PGD_MASKED_BITS)
60 #define pgd_page(pgd) virt_to_page(pgd_page_vaddr(pgd))
[all...]
/arch/arm/mm/
H A Dpgd.c2 * linux/arch/arm/mm/pgd.c
24 #define __pgd_free(pgd) kfree(pgd)
27 #define __pgd_free(pgd) free_pages((unsigned long)pgd, 2)
109 pgd_t *pgd; local
117 pgd = pgd_base + pgd_index(0);
118 if (pgd_none_or_clear_bad(pgd))
121 pud = pud_offset(pgd, 0);
136 pgd_clear(pgd);
[all...]
/arch/arm/plat-samsung/include/plat/
H A Dsysmmu.h35 * #pgd: Base physical address of the 1st level page table
40 void s5p_sysmmu_enable(sysmmu_ips ips, unsigned long pgd);
54 * @pgd: The page table base address.
60 void s5p_sysmmu_set_tablebase_pgd(sysmmu_ips ips, unsigned long pgd);
89 #define s5p_sysmmu_enable(ips, pgd) do { } while (0)
91 #define s5p_sysmmu_set_tablebase_pgd(ips, pgd) do { } while (0)
/arch/mn10300/mm/
H A Dpgtable.c37 pgd_t *pgd; local
49 pgd = swapper_pg_dir + pgd_index(vaddr);
50 if (pgd_none(*pgd)) {
54 pud = pud_offset(pgd, vaddr);
87 * List of all pgd's needed for non-PAE so it can invalidate entries
88 * in both cached and uncached pgd's; not needed for PAE since the
103 static inline void pgd_list_add(pgd_t *pgd) argument
105 struct page *page = virt_to_page(pgd);
113 static inline void pgd_list_del(pgd_t *pgd) argument
115 struct page *next, **pprev, *page = virt_to_page(pgd);
123 pgd_ctor(void *pgd) argument
143 pgd_dtor(void *pgd) argument
157 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
[all...]
/arch/hexagon/include/asm/
H A Dpgalloc.h36 pgd_t *pgd; local
38 pgd = (pgd_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
48 memcpy(pgd, swapper_pg_dir, PTRS_PER_PGD*sizeof(pgd_t *));
52 mm->context.ptbase = __pa(pgd);
54 return pgd;
57 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
59 free_page((unsigned long) pgd);
133 pmdindex = (pgd_t *)pmd - mm->pgd;
134 ppmd = (pmd_t *)current->active_mm->pgd + pmdindex;
/arch/frv/mm/
H A Dpgalloc.c69 * List of all pgd's needed for non-PAE so it can invalidate entries
70 * in both cached and uncached pgd's; not needed for PAE since the
85 static inline void pgd_list_add(pgd_t *pgd) argument
87 struct page *page = virt_to_page(pgd);
95 static inline void pgd_list_del(pgd_t *pgd) argument
97 struct page *next, **pprev, *page = virt_to_page(pgd);
105 void pgd_ctor(void *pgd) argument
112 memcpy((pgd_t *) pgd + USER_PGDS_IN_LAST_PML4,
119 pgd_list_add(pgd);
121 memset(pgd,
125 pgd_dtor(void *pgd) argument
139 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
[all...]
/arch/x86/mm/
H A Dpgtable.c73 static inline void pgd_list_add(pgd_t *pgd) argument
75 struct page *page = virt_to_page(pgd);
80 static inline void pgd_list_del(pgd_t *pgd) argument
82 struct page *page = virt_to_page(pgd);
91 static void pgd_set_mm(pgd_t *pgd, struct mm_struct *mm) argument
93 BUILD_BUG_ON(sizeof(virt_to_page(pgd)->index) < sizeof(mm));
94 virt_to_page(pgd)->index = (pgoff_t)mm;
102 static void pgd_ctor(struct mm_struct *mm, pgd_t *pgd) argument
104 /* If the pgd points to a shared pagetable level (either the
110 clone_pgd_range(pgd
122 pgd_dtor(pgd_t *pgd) argument
220 pgd_t pgd = pgdp[i]; local
233 pgd_prepopulate_pmd(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmds[]) argument
258 pgd_t *pgd; local
296 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
[all...]
H A Dpgtable_32.c29 pgd_t *pgd; local
34 pgd = swapper_pg_dir + pgd_index(vaddr);
35 if (pgd_none(*pgd)) {
39 pud = pud_offset(pgd, vaddr);
70 pgd_t *pgd; local
82 pgd = swapper_pg_dir + pgd_index(vaddr);
83 if (pgd_none(*pgd)) {
87 pud = pud_offset(pgd, vaddr);
/arch/parisc/include/asm/
H A Dpgalloc.h12 /* Allocate the top level pgd (page directory)
15 * allocate the first pmd adjacent to the pgd. This means that we can
16 * subtract a constant offset to get to it. The pmd and pgd sizes are
23 pgd_t *pgd = (pgd_t *)__get_free_pages(GFP_KERNEL, local
25 pgd_t *actual_pgd = pgd;
27 if (likely(pgd != NULL)) {
28 memset(pgd, 0, PAGE_SIZE<<PGD_ALLOC_ORDER);
37 + (__u32)(__pa((unsigned long)pgd) >> PxD_VALUE_SHIFT));
40 __pgd_val_set(*pgd, PxD_FLAG_ATTACHED);
46 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
58 pgd_populate(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmd) argument
[all...]
/arch/sh/mm/
H A Dhugetlbpage.c27 pgd_t *pgd; local
32 pgd = pgd_offset(mm, addr);
33 if (pgd) {
34 pud = pud_alloc(mm, pgd, addr);
47 pgd_t *pgd; local
52 pgd = pgd_offset(mm, addr);
53 if (pgd) {
54 pud = pud_offset(pgd, addr);
/arch/unicore32/include/asm/
H A Dcpu-single.h21 #define cpu_switch_mm(pgd, mm) cpu_do_switch_mm(virt_to_phys(pgd), mm)
/arch/m32r/include/asm/
H A Dpgtable-2level.h26 printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
30 * setup: the pgd is never bad, and a pmd always exists (as it's folded
31 * into the pgd entry)
33 static inline int pgd_none(pgd_t pgd) { return 0; } argument
34 static inline int pgd_bad(pgd_t pgd) { return 0; } argument
35 static inline int pgd_present(pgd_t pgd) { return 1; } argument
53 #define pgd_page_vaddr(pgd) \
54 ((unsigned long) __va(pgd_val(pgd) & PAGE_MASK))
57 #define pgd_page(pgd) (mem_map + ((pgd_val(pgd) >> PAGE_SHIF
[all...]
H A Dpgalloc.h23 pgd_t *pgd = (pgd_t *)__get_free_page(GFP_KERNEL|__GFP_ZERO); local
25 return pgd;
28 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
30 free_page((unsigned long)pgd);
65 * inside the pgd, so has no extra memory associated with it.
66 * (In the PAE case we free the pmds as part of the pgd.)
/arch/mips/mm/
H A Dhugetlbpage.c28 pgd_t *pgd; local
32 pgd = pgd_offset(mm, addr);
33 pud = pud_alloc(mm, pgd, addr);
42 pgd_t *pgd; local
46 pgd = pgd_offset(mm, addr);
47 if (pgd_present(*pgd)) {
48 pud = pud_offset(pgd, addr);
/arch/ia64/xen/
H A Dxencomm.c51 pgd_t *pgd; local
65 pgd = pgd_offset_k(vaddr);
66 if (pgd_none(*pgd) || pgd_bad(*pgd))
69 pud = pud_offset(pgd, vaddr);
/arch/unicore32/kernel/
H A Dhibernate.c35 static pmd_t *resume_one_md_table_init(pgd_t *pgd) argument
40 pud = pud_offset(pgd, 0);
75 pgd_t *pgd; local
81 pgd = pgd_base + pgd_idx;
84 for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) {
85 pmd = resume_one_md_table_init(pgd);
/arch/sh/include/asm/
H A Dmmu_context_32.h50 static inline void set_TTB(pgd_t *pgd) argument
52 __raw_writel((unsigned long)pgd, MMU_TTB);
H A Dmmu_context_64.h75 #define set_TTB(pgd) (mmu_pdtp_cache = (pgd))
/arch/unicore32/mm/
H A Dmm.h20 static inline pmd_t *pmd_off(pgd_t *pgd, unsigned long virt) argument
22 return pmd_offset((pud_t *)pgd, virt);
/arch/x86/include/asm/
H A Dmmzone_32.h17 extern void resume_map_numa_kva(pgd_t *pgd);
21 static inline void resume_map_numa_kva(pgd_t *pgd) {} argument
/arch/x86/power/
H A Dhibernate_32.c35 static pmd_t *resume_one_md_table_init(pgd_t *pgd) argument
45 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT));
46 pud = pud_offset(pgd, 0);
50 pud = pud_offset(pgd, 0);
86 pgd_t *pgd; local
92 pgd = pgd_base + pgd_idx;
95 for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) {
96 pmd = resume_one_md_table_init(pgd);
/arch/um/kernel/
H A Dmem.c118 pgd_t *pgd; local
127 pgd = pgd_base + i;
129 for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) {
130 pud = pud_offset(pgd, vaddr);
163 pgd_t *pgd; local
175 pgd = swapper_pg_dir + pgd_index(vaddr);
176 pud = pud_offset(pgd, vaddr);
189 pgd_t *pgd; local
205 pgd = swapper_pg_dir + pgd_index(vaddr);
206 pud = pud_offset(pgd, vadd
273 pgd_t *pgd = (pgd_t *)__get_free_page(GFP_KERNEL); local
284 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
[all...]
/arch/tile/mm/
H A Dpgtable.c92 pgd_t *pgd; local
97 pgd = swapper_pg_dir + pgd_index(vaddr);
98 if (pgd_none(*pgd)) {
102 pud = pud_offset(pgd, vaddr);
160 pgd_t *pgd; local
172 pgd = swapper_pg_dir + pgd_index(addr);
173 pud = pud_offset(pgd, addr);
192 /* Walk every pgd on the system and update the pmd there. */
196 pgd = list_to_pgd(pos) + pgd_index(addr);
197 pud = pud_offset(pgd, add
229 pgd_list_add(pgd_t *pgd) argument
234 pgd_list_del(pgd_t *pgd) argument
242 pgd_ctor(pgd_t *pgd) argument
266 pgd_dtor(pgd_t *pgd) argument
277 pgd_t *pgd = kmem_cache_alloc(pgd_cache, GFP_KERNEL); local
283 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
395 pgd_t *pgd; local
[all...]
/arch/alpha/include/asm/
H A Dpgalloc.h27 pgd_populate(struct mm_struct *mm, pgd_t *pgd, pmd_t *pmd) argument
29 pgd_set(pgd, pmd);
35 pgd_free(struct mm_struct *mm, pgd_t *pgd) argument
37 free_page((unsigned long)pgd);
/arch/mips/include/asm/mach-powertv/
H A Ddma-coherence.h26 pgd_t *pgd; local
35 pgd = pgd_offset_k(virt_addr);
37 if (!pgd_none(*pgd)) {
39 pud = pud_offset(pgd, virt_addr);

Completed in 322 milliseconds

1234567891011