/mm/ |
H A D | oom_kill.c | 51 * @mask: nodemask passed to page allocator for mempolicy ooms 122 const struct mem_cgroup *memcg, const nodemask_t *nodemask) 133 /* p may not have freeable memory in nodemask */ 134 if (!has_intersects_mems_allowed(p, nodemask)) 150 const nodemask_t *nodemask, unsigned long totalpages) 155 if (oom_unkillable_task(p, memcg, nodemask)) 199 gfp_t gfp_mask, nodemask_t *nodemask, 222 * This is not a __GFP_THISNODE allocation, so a truncated nodemask in 226 if (nodemask && !nodes_subset(node_states[N_MEMORY], *nodemask)) { 121 oom_unkillable_task(struct task_struct *p, const struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 149 oom_badness(struct task_struct *p, struct mem_cgroup *memcg, const nodemask_t *nodemask, unsigned long totalpages) argument 198 constrained_alloc(struct zonelist *zonelist, gfp_t gfp_mask, nodemask_t *nodemask, unsigned long *totalpages) argument 248 constrained_alloc(struct zonelist *zonelist, gfp_t gfp_mask, nodemask_t *nodemask, unsigned long *totalpages) argument 257 oom_scan_process_thread(struct task_struct *task, unsigned long totalpages, const nodemask_t *nodemask, bool force_kill) argument 301 select_bad_process(unsigned int *ppoints, unsigned long totalpages, const nodemask_t *nodemask, bool force_kill) argument 356 dump_tasks(const struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 388 dump_header(struct task_struct *p, gfp_t gfp_mask, int order, struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 429 oom_kill_process(struct task_struct *p, gfp_t gfp_mask, int order, unsigned int points, unsigned long totalpages, struct mem_cgroup *memcg, nodemask_t *nodemask, const char *message) argument 539 check_panic_on_oom(enum oom_constraint constraint, gfp_t gfp_mask, int order, const nodemask_t *nodemask) argument 631 out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order, nodemask_t *nodemask, bool force_kill) argument [all...] |
H A D | page_alloc.c | 41 #include <linux/nodemask.h> 1949 get_page_from_freelist(gfp_t gfp_mask, nodemask_t *nodemask, unsigned int order, argument 1972 high_zoneidx, nodemask) { 2247 nodemask_t *nodemask, struct zone *preferred_zone, 2271 page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, 2296 out_of_memory(zonelist, gfp_mask, order, nodemask, false); 2308 nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, 2321 nodemask, mode, 2346 page = get_page_from_freelist(gfp_mask, nodemask, 2382 nodemask_t *nodemask, in 2245 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, struct zone *preferred_zone, int classzone_idx, int migratetype) argument 2306 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, int classzone_idx, int migratetype, enum migrate_mode mode, int *contended_compaction, bool *deferred_compaction) argument 2380 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, int classzone_idx, int migratetype, enum migrate_mode mode, int *contended_compaction, bool *deferred_compaction) argument 2392 __perform_reclaim(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, nodemask_t *nodemask) argument 2420 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, int classzone_idx, int migratetype, unsigned long *did_some_progress) argument 2462 __alloc_pages_high_priority(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, struct zone *preferred_zone, int classzone_idx, int migratetype) argument 2481 wake_all_kswapds(unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, struct zone *preferred_zone, nodemask_t *nodemask) argument 2550 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, struct zone *preferred_zone, int classzone_idx, int migratetype) argument 2795 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, nodemask_t *nodemask) argument [all...] |
H A D | vmscan.c | 76 nodemask_t *nodemask; member in struct:scan_control 2443 gfp_zone(sc->gfp_mask), sc->nodemask) { 2648 nodemask_t *nodemask) 2686 gfp_mask, nodemask) { 2732 gfp_t gfp_mask, nodemask_t *nodemask) 2739 .nodemask = nodemask, 2751 if (throttle_direct_reclaim(gfp_mask, zonelist, nodemask)) 2647 throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, nodemask_t *nodemask) argument 2731 try_to_free_pages(struct zonelist *zonelist, int order, gfp_t gfp_mask, nodemask_t *nodemask) argument
|
H A D | mempolicy.c | 76 #include <linux/nodemask.h> 106 #define MPOL_MF_INVERT (MPOL_MF_INTERNAL << 1) /* Invert check for nodemask */ 165 /* Check that the nodemask contains at least one populated zone */ 166 static int is_valid_nodemask(const nodemask_t *nodemask) argument 168 return nodes_intersects(*nodemask, node_states[N_MEMORY]); 212 * mpol_set_nodemask is called after mpol_new() to set up the nodemask, if 215 * handle an empty nodemask with MPOL_PREFERRED here. 276 * MPOL_F_RELATIVE_NODES if the nodemask is empty (local allocation). 277 * All other modes require a valid pointer to a non-empty nodemask. 444 * Rebind each vma in mm to new nodemask 1860 huge_zonelist(struct vm_area_struct *vma, unsigned long addr, gfp_t gfp_flags, struct mempolicy **mpol, nodemask_t **nodemask) argument [all...] |
H A D | compaction.c | 1287 * @nodemask: The allowed nodes to allocate from 1296 int order, gfp_t gfp_mask, nodemask_t *nodemask, 1321 nodemask) { 1295 try_to_compact_pages(struct zonelist *zonelist, int order, gfp_t gfp_mask, nodemask_t *nodemask, enum migrate_mode mode, int *contended, struct zone **candidate_zone) argument
|
H A D | hugetlb.c | 13 #include <linux/nodemask.h> 559 nodemask_t *nodemask; local 581 htlb_alloc_mask(h), &mpol, &nodemask); 584 MAX_NR_ZONES - 1, nodemask) {
|
H A D | kmemleak.c | 92 #include <linux/nodemask.h>
|
H A D | slab.c | 108 #include <linux/nodemask.h>
|