/mm/ |
H A D | oom_kill.c | 92 * @mask: nodemask passed to page allocator for mempolicy ooms 155 const struct mem_cgroup *memcg, const nodemask_t *nodemask) 166 /* p may not have freeable memory in nodemask */ 167 if (!has_intersects_mems_allowed(p, nodemask)) 183 const nodemask_t *nodemask, unsigned long totalpages) 187 if (oom_unkillable_task(p, memcg, nodemask)) 246 gfp_t gfp_mask, nodemask_t *nodemask, 269 * This is not a __GFP_THISNODE allocation, so a truncated nodemask in 273 if (nodemask && !nodes_subset(node_states[N_HIGH_MEMORY], *nodemask)) { 154 oom_unkillable_task(struct task_struct *p, const struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 182 oom_badness(struct task_struct *p, struct mem_cgroup *memcg, const nodemask_t *nodemask, unsigned long totalpages) argument 245 constrained_alloc(struct zonelist *zonelist, gfp_t gfp_mask, nodemask_t *nodemask, unsigned long *totalpages) argument 295 constrained_alloc(struct zonelist *zonelist, gfp_t gfp_mask, nodemask_t *nodemask, unsigned long *totalpages) argument 310 select_bad_process(unsigned int *ppoints, unsigned long totalpages, struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 390 dump_tasks(const struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 419 dump_header(struct task_struct *p, gfp_t gfp_mask, int order, struct mem_cgroup *memcg, const nodemask_t *nodemask) argument 485 oom_kill_process(struct task_struct *p, gfp_t gfp_mask, int order, unsigned int points, unsigned long totalpages, struct mem_cgroup *memcg, nodemask_t *nodemask, const char *message) argument 542 check_panic_on_oom(enum oom_constraint constraint, gfp_t gfp_mask, int order, const nodemask_t *nodemask) argument 709 out_of_memory(struct zonelist *zonelist, gfp_t gfp_mask, int order, nodemask_t *nodemask) argument [all...] |
H A D | compaction.c | 634 * @nodemask: The allowed nodes to allocate from 640 int order, gfp_t gfp_mask, nodemask_t *nodemask, 662 nodemask) { 639 try_to_compact_pages(struct zonelist *zonelist, int order, gfp_t gfp_mask, nodemask_t *nodemask, bool sync) argument
|
H A D | mempolicy.c | 74 #include <linux/nodemask.h> 102 #define MPOL_MF_INVERT (MPOL_MF_INTERNAL << 1) /* Invert check for nodemask */ 140 /* Check that the nodemask contains at least one populated zone */ 141 static int is_valid_nodemask(const nodemask_t *nodemask) argument 145 for_each_node_mask(nd, *nodemask) { 199 * mpol_set_nodemask is called after mpol_new() to set up the nodemask, if 202 * handle an empty nodemask with MPOL_PREFERRED here. 263 * MPOL_F_RELATIVE_NODES if the nodemask is empty (local allocation). 264 * All other modes require a valid pointer to a non-empty nodemask. 427 * Rebind each vma in mm to new nodemask 1690 huge_zonelist(struct vm_area_struct *vma, unsigned long addr, gfp_t gfp_flags, struct mempolicy **mpol, nodemask_t **nodemask) argument [all...] |
H A D | hugetlb.c | 13 #include <linux/nodemask.h> 459 nodemask_t *nodemask; local 466 htlb_alloc_mask, &mpol, &nodemask); 481 MAX_NR_ZONES - 1, nodemask) {
|
H A D | vmscan.c | 115 nodemask_t *nodemask; member in struct:scan_control 2239 gfp_zone(sc->gfp_mask), sc->nodemask) { 2300 gfp_zone(sc->gfp_mask), sc->nodemask) { 2429 gfp_t gfp_mask, nodemask_t *nodemask) 2440 .nodemask = nodemask, 2514 .nodemask = NULL, /* we don't care the placement */ 2428 try_to_free_pages(struct zonelist *zonelist, int order, gfp_t gfp_mask, nodemask_t *nodemask) argument
|
H A D | page_alloc.c | 41 #include <linux/nodemask.h> 1704 get_page_from_freelist(gfp_t gfp_mask, nodemask_t *nodemask, unsigned int order, argument 1723 high_zoneidx, nodemask) { 1931 nodemask_t *nodemask, struct zone *preferred_zone, 1947 page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, 1972 out_of_memory(zonelist, gfp_mask, order, nodemask); 1984 nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, 2001 nodemask, sync_migration); 2009 page = get_page_from_freelist(gfp_mask, nodemask, 2043 nodemask_t *nodemask, in 1929 __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, struct zone *preferred_zone, int migratetype) argument 1982 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, int migratetype, bool sync_migration, bool *deferred_compaction, unsigned long *did_some_progress) argument 2041 __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, int migratetype, bool sync_migration, bool *deferred_compaction, unsigned long *did_some_progress) argument 2054 __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, int migratetype, unsigned long *did_some_progress) argument 2111 __alloc_pages_high_priority(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, struct zone *preferred_zone, int migratetype) argument 2185 __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, nodemask_t *nodemask, struct zone *preferred_zone, int migratetype) argument 2377 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, nodemask_t *nodemask) argument [all...] |