/include/asm-generic/ |
H A D | tlbflush.h | 14 static inline void flush_tlb_mm(struct mm_struct *mm) argument
|
H A D | mm_hooks.h | 10 struct mm_struct *mm) 14 static inline void arch_exit_mmap(struct mm_struct *mm) argument 9 arch_dup_mmap(struct mm_struct *oldmm, struct mm_struct *mm) argument
|
H A D | mmu_context.h | 14 static inline void enter_lazy_tlb(struct mm_struct *mm, argument 20 struct mm_struct *mm) 25 static inline void destroy_context(struct mm_struct *mm) argument 30 struct mm_struct *mm) 19 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 29 deactivate_mm(struct task_struct *task, struct mm_struct *mm) argument
|
H A D | pgtable-nopmd.h | 35 #define pud_populate(mm, pmd, pte) do { } while (0) 58 #define pmd_alloc_one(mm, address) NULL 59 static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd) argument
|
H A D | tlb.h | 6 * Based on code from mm/memory.c Copyright Linus Torvalds and others. 81 /* struct mmu_gather is an opaque type used by the mm code for passing around 85 struct mm_struct *mm; member in struct:mmu_gather 114 void tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, bool fullmm); 176 #define tlb_migrate_finish(mm) do {} while (0)
|
H A D | pgtable.h | 73 static inline pte_t ptep_get_and_clear(struct mm_struct *mm, argument 78 pte_clear(mm, address, ptep); 85 static inline pmd_t pmdp_get_and_clear(struct mm_struct *mm, argument 90 pmd_clear(mm, address, pmdp); 97 static inline pte_t ptep_get_and_clear_full(struct mm_struct *mm, argument 102 pte = ptep_get_and_clear(mm, address, ptep); 113 static inline void pte_clear_not_present_full(struct mm_struct *mm, argument 118 pte_clear(mm, address, ptep); 136 static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long address, pte_t *ptep) argument 139 set_pte_at(mm, addres 145 pmdp_set_wrprotect(struct mm_struct *mm, unsigned long address, pmd_t *pmdp) argument 152 pmdp_set_wrprotect(struct mm_struct *mm, unsigned long address, pmd_t *pmdp) argument 289 __ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 301 __ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte) argument 327 ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, pte_t *ptep) argument 338 ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte) argument [all...] |
/include/linux/ |
H A D | khugepaged.h | 7 extern int __khugepaged_enter(struct mm_struct *mm); 8 extern void __khugepaged_exit(struct mm_struct *mm); 25 static inline int khugepaged_fork(struct mm_struct *mm, struct mm_struct *oldmm) argument 28 return __khugepaged_enter(mm); 32 static inline void khugepaged_exit(struct mm_struct *mm) argument 34 if (test_bit(MMF_VM_HUGEPAGE, &mm->flags)) 35 __khugepaged_exit(mm); 50 static inline int khugepaged_fork(struct mm_struct *mm, struct mm_struct *oldmm) argument 54 static inline void khugepaged_exit(struct mm_struct *mm) argument
|
H A D | rslib.h | 27 * @mm: Bits per symbol 28 * @nn: Symbols per block (= (1<<mm)-1) 42 int mm; member in struct:rs_control 94 * rs->mm = number of bits per symbol 95 * rs->nn = (2^rs->mm) - 1 104 x = (x >> rs->mm) + (x & rs->nn);
|
H A D | migrate.h | 4 #include <linux/mm.h> 27 extern int migrate_vmas(struct mm_struct *mm, 46 static inline int migrate_vmas(struct mm_struct *mm, argument
|
H A D | ksm.h | 11 #include <linux/mm.h> 25 int __ksm_enter(struct mm_struct *mm); 26 void __ksm_exit(struct mm_struct *mm); 28 static inline int ksm_fork(struct mm_struct *mm, struct mm_struct *oldmm) argument 31 return __ksm_enter(mm); 35 static inline void ksm_exit(struct mm_struct *mm) argument 37 if (test_bit(MMF_VM_MERGEABLE, &mm->flags)) 38 __ksm_exit(mm); 95 static inline int ksm_fork(struct mm_struct *mm, struct mm_struct *oldmm) argument 100 static inline void ksm_exit(struct mm_struct *mm) argument [all...] |
H A D | mmu_notifier.h | 15 * mm->mmu_notifier_mm inside the mm_take_all_locks() protected 20 /* all mmu notifiers registerd in this mm are queued in this list */ 28 * Called either by mmu_notifier_unregister or when the mm is 31 * methods (the ones invoked outside the mm context) and it 36 * tsk->mm == mm exits. 43 * last thread of this mm quits, you've also to be sure that 51 struct mm_struct *mm); 61 struct mm_struct *mm, 71 struct mm_struct *mm, 161 mm_has_notifiers(struct mm_struct *mm) argument 187 mmu_notifier_release(struct mm_struct *mm) argument 193 mmu_notifier_clear_flush_young(struct mm_struct *mm, unsigned long address) argument 201 mmu_notifier_test_young(struct mm_struct *mm, unsigned long address) argument 209 mmu_notifier_change_pte(struct mm_struct *mm, unsigned long address, pte_t pte) argument 216 mmu_notifier_invalidate_page(struct mm_struct *mm, unsigned long address) argument 223 mmu_notifier_invalidate_range_start(struct mm_struct *mm, unsigned long start, unsigned long end) argument 230 mmu_notifier_invalidate_range_end(struct mm_struct *mm, unsigned long start, unsigned long end) argument 237 mmu_notifier_mm_init(struct mm_struct *mm) argument 242 mmu_notifier_mm_destroy(struct mm_struct *mm) argument 326 mmu_notifier_release(struct mm_struct *mm) argument 330 mmu_notifier_clear_flush_young(struct mm_struct *mm, unsigned long address) argument 336 mmu_notifier_test_young(struct mm_struct *mm, unsigned long address) argument 342 mmu_notifier_change_pte(struct mm_struct *mm, unsigned long address, pte_t pte) argument 347 mmu_notifier_invalidate_page(struct mm_struct *mm, unsigned long address) argument 352 mmu_notifier_invalidate_range_start(struct mm_struct *mm, unsigned long start, unsigned long end) argument 357 mmu_notifier_invalidate_range_end(struct mm_struct *mm, unsigned long start, unsigned long end) argument 362 mmu_notifier_mm_init(struct mm_struct *mm) argument 366 mmu_notifier_mm_destroy(struct mm_struct *mm) argument [all...] |
H A D | swapops.h | 138 extern void migration_entry_wait(struct mm_struct *mm, pmd_t *pmd, 149 static inline void migration_entry_wait(struct mm_struct *mm, pmd_t *pmd, argument
|
H A D | binfmts.h | 37 struct mm_struct *mm; member in struct:linux_binprm
|
H A D | rmap.h | 4 * Declarations for Reverse Mapping functions in mm/rmap.c 9 #include <linux/mm.h> 157 * Called from mm/vmscan.c to handle paging out 183 * Called from mm/filemap_xip.c to unmap empty zero page 188 static inline pte_t *page_check_address(struct page *page, struct mm_struct *mm, argument 194 __cond_lock(*ptlp, ptep = __page_check_address(page, mm, address,
|
H A D | futex.h | 146 * (no reference on an inode or mm) 150 * (but private mapping on an mm, and reference taken on it) 154 #define FUT_OFF_MMSHARED 2 /* We set bit 1 if key has a reference on mm */ 164 struct mm_struct *mm; member in struct:futex_key::__anon415
|
H A D | memcontrol.h | 53 extern int mem_cgroup_newpage_charge(struct page *page, struct mm_struct *mm, 56 extern int mem_cgroup_try_charge_swapin(struct mm_struct *mm, 62 extern int mem_cgroup_cache_charge(struct page *page, struct mm_struct *mm, 86 extern struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm); 92 int mm_match_cgroup(const struct mm_struct *mm, const struct mem_cgroup *cgroup) argument 96 memcg = mem_cgroup_from_task(rcu_dereference((mm)->owner)); 193 void mem_cgroup_count_vm_event(struct mm_struct *mm, enum vm_event_item idx); 206 struct mm_struct *mm, gfp_t gfp_mask) 212 struct mm_struct *mm, gfp_t gfp_mask) 217 static inline int mem_cgroup_try_charge_swapin(struct mm_struct *mm, argument 205 mem_cgroup_newpage_charge(struct page *page, struct mm_struct *mm, gfp_t gfp_mask) argument 211 mem_cgroup_cache_charge(struct page *page, struct mm_struct *mm, gfp_t gfp_mask) argument 282 try_get_mem_cgroup_from_mm(struct mm_struct *mm) argument 287 mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) argument 411 mem_cgroup_count_vm_event(struct mm_struct *mm, enum vm_event_item idx) argument [all...] |
H A D | pagemap.h | 7 #include <linux/mm.h> 305 extern int __lock_page_or_retry(struct page *page, struct mm_struct *mm, 351 static inline int lock_page_or_retry(struct page *page, struct mm_struct *mm, argument 355 return trylock_page(page) || __lock_page_or_retry(page, mm, flags);
|
H A D | aio.h | 183 struct mm_struct *mm; member in struct:kioctx 216 extern void exit_aio(struct mm_struct *mm); 225 static inline void exit_aio(struct mm_struct *mm) { } argument
|
H A D | mempolicy.h | 177 * Maintain the policies in a pseudo mm that contains vmas. The vmas 178 * carry the policy. As a special twist the pseudo mm is indexed in pages, not 209 extern void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new); 228 int do_migrate_pages(struct mm_struct *mm, 329 static inline void mpol_rebind_mm(struct mm_struct *mm, nodemask_t *new) argument 357 static inline int do_migrate_pages(struct mm_struct *mm, argument
|
H A D | mm_types.h | 220 unsigned long vm_flags; /* Flags, see mm.h. */ 306 void (*unmap_area) (struct mm_struct *mm, unsigned long addr); 320 struct list_head mmlist; /* List of maybe swapped mm's. These are globally strung 362 * Look at mm/thrash.c 378 * user/owner of this mm. All of the following must be true in 381 * current == mm->owner 382 * current->mm != mm 383 * new_owner->mm == mm 403 mm_init_cpumask(struct mm_struct *mm) argument 411 mm_cpumask(struct mm_struct *mm) argument [all...] |
H A D | swap.h | 210 /* linux/mm/page_alloc.c */ 221 /* linux/mm/swap.c */ 251 /* linux/mm/vmscan.c */ 316 /* linux/mm/page_io.c */ 321 /* linux/mm/swap_state.c */ 337 /* linux/mm/swapfile.c */ 358 /* linux/mm/thrash.c */ 364 static inline int has_swap_token(struct mm_struct *mm) argument 366 return (mm == swap_token_mm); 369 static inline void put_swap_token(struct mm_struct *mm) argument 480 put_swap_token(struct mm_struct *mm) argument 484 grab_swap_token(struct mm_struct *mm) argument 488 has_swap_token(struct mm_struct *mm) argument [all...] |
H A D | kvm_host.h | 17 #include <linux/mm.h> 98 struct mm_struct *mm; member in struct:kvm_async_pf 247 struct mm_struct *mm; /* userspace tied to this vm */ member in struct:kvm
|
H A D | mm.h | 54 * The idea being to have a "virtual" mm in the same way 56 * mm details, and allowing different kinds of memory mappings 144 * Note: mm/huge_memory.c VM_NO_THP depends on this definition. 165 * mm use. Note also that this check will not work if the pfn mapping is 232 * in mm/mempolicy.c will do this automatically. 424 * (see mm/page_alloc.c). 528 * page, and the fields in 'struct page' are the responsibility of mm/slab.c 929 struct mm_struct *mm; member in struct:mm_walk 965 extern int handle_mm_fault(struct mm_struct *mm, struct vm_area_struct *vma, 967 extern int fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm, 970 handle_mm_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, unsigned int flags) argument 978 fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm, unsigned long address, unsigned int fault_flags) argument 1066 get_mm_counter(struct mm_struct *mm, int member) argument 1081 add_mm_counter(struct mm_struct *mm, int member, long value) argument 1086 inc_mm_counter(struct mm_struct *mm, int member) argument 1091 dec_mm_counter(struct mm_struct *mm, int member) argument 1096 get_mm_rss(struct mm_struct *mm) argument 1102 get_mm_hiwater_rss(struct mm_struct *mm) argument 1107 get_mm_hiwater_vm(struct mm_struct *mm) argument 1112 update_hiwater_rss(struct mm_struct *mm) argument 1120 update_hiwater_vm(struct mm_struct *mm) argument 1126 setmax_mm_hiwater_rss(unsigned long *maxrss, struct mm_struct *mm) argument 1138 sync_mm_rss(struct mm_struct *mm) argument 1147 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument 1156 __pud_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) argument 1166 __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) argument 1184 pud_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) argument 1190 pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) argument 1468 find_vma_intersection(struct mm_struct * mm, unsigned long start_addr, unsigned long end_addr) argument 1483 find_exact_vma(struct mm_struct *mm, unsigned long vm_start, unsigned long vm_end) argument 1533 vm_stat_account(struct mm_struct *mm, unsigned long flags, struct file *file, long pages) argument [all...] |
/include/rdma/ |
H A D | ib_umem.h | 51 struct mm_struct *mm; member in struct:ib_umem
|
/include/drm/ |
H A D | drm_mm.h | 55 struct drm_mm *mm; member in struct:drm_mm_node 83 static inline bool drm_mm_initialized(struct drm_mm *mm) argument 85 return mm->hole_stack.next; 87 #define drm_mm_for_each_node(entry, mm) list_for_each_entry(entry, \ 88 &(mm)->head_node.node_list, \ 90 #define drm_mm_for_each_scanned_node_reverse(entry, n, mm) \ 91 for (entry = (mm)->prev_scanned_node, \ 143 extern int drm_mm_insert_node(struct drm_mm *mm, struct drm_mm_node *node, 145 extern int drm_mm_insert_node_in_range(struct drm_mm *mm, 152 extern struct drm_mm_node *drm_mm_search_free(const struct drm_mm *mm, [all...] |