/mm/ |
H A D | interval_tree.c | 31 struct rb_root *root) 58 rb_insert_augmented(&node->shared.linear.rb, root, 77 struct rb_root *root) 83 __anon_vma_interval_tree_insert(node, root); 87 struct rb_root *root) 89 __anon_vma_interval_tree_remove(node, root); 93 anon_vma_interval_tree_iter_first(struct rb_root *root, argument 96 return __anon_vma_interval_tree_iter_first(root, first, last); 29 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root *root) argument 76 anon_vma_interval_tree_insert(struct anon_vma_chain *node, struct rb_root *root) argument 86 anon_vma_interval_tree_remove(struct anon_vma_chain *node, struct rb_root *root) argument
|
H A D | cleancache.c | 393 struct dentry *root = debugfs_create_dir("cleancache", NULL); local 394 if (root == NULL) 396 debugfs_create_u64("succ_gets", S_IRUGO, root, &cleancache_succ_gets); 398 root, &cleancache_failed_gets); 399 debugfs_create_u64("puts", S_IRUGO, root, &cleancache_puts); 401 root, &cleancache_invalidates);
|
H A D | frontswap.c | 444 struct dentry *root = debugfs_create_dir("frontswap", NULL); local 445 if (root == NULL) 447 debugfs_create_u64("loads", S_IRUGO, root, &frontswap_loads); 448 debugfs_create_u64("succ_stores", S_IRUGO, root, &frontswap_succ_stores); 449 debugfs_create_u64("failed_stores", S_IRUGO, root, 452 root, &frontswap_invalidates);
|
H A D | sparse.c | 82 unsigned long root = SECTION_NR_TO_ROOT(section_nr); local 85 if (mem_section[root]) 92 mem_section[root] = section; 111 struct mem_section* root; local 114 root = __nr_to_section(root_nr * SECTIONS_PER_ROOT); 115 if (!root) 118 if ((ms >= root) && (ms < (root + SECTIONS_PER_ROOT))) 124 return (root_nr * SECTIONS_PER_ROOT) + (ms - root);
|
H A D | zswap.c | 239 static struct zswap_entry *zswap_rb_search(struct rb_root *root, pgoff_t offset) argument 241 struct rb_node *node = root->rb_node; 260 static int zswap_rb_insert(struct rb_root *root, struct zswap_entry *entry, argument 263 struct rb_node **link = &root->rb_node, *parent = NULL; 279 rb_insert_color(&entry->rbnode, root); 283 static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry) argument 286 rb_erase(&entry->rbnode, root); 325 static struct zswap_entry *zswap_entry_find_get(struct rb_root *root, argument 330 entry = zswap_rb_search(root, offset);
|
H A D | memblock.c | 1581 struct dentry *root = debugfs_create_dir("memblock", NULL); local 1582 if (!root) 1584 debugfs_create_file("memory", S_IRUGO, root, &memblock.memory, &memblock_debug_fops); 1585 debugfs_create_file("reserved", S_IRUGO, root, &memblock.reserved, &memblock_debug_fops); 1587 debugfs_create_file("physmem", S_IRUGO, root, &memblock.physmem, &memblock_debug_fops);
|
H A D | rmap.c | 76 * Initialise the anon_vma root to point to itself. If called 77 * from fork, the root will be reset to the parents anon_vma. 79 anon_vma->root = anon_vma; 107 if (rwsem_is_locked(&anon_vma->root->rwsem)) { 211 * This is a useful helper function for locking the anon_vma root as 215 * Such anon_vma's should have the same root, so you'd expect to see 218 static inline struct anon_vma *lock_anon_vma_root(struct anon_vma *root, struct anon_vma *anon_vma) argument 220 struct anon_vma *new_root = anon_vma->root; 221 if (new_root != root) { 222 if (WARN_ON_ONCE(root)) 230 unlock_anon_vma_root(struct anon_vma *root) argument 243 struct anon_vma *root = NULL; local 328 struct anon_vma *root = NULL; local 1592 struct anon_vma *root = anon_vma->root; local [all...] |
H A D | ksm.c | 1154 struct rb_root *root; local 1168 root = root_stable_tree + nid; 1170 new = &root->rb_node; 1225 rb_insert_color(&page_node->node, root); 1233 rb_replace_node(&stable_node->node, &page_node->node, root); 1236 rb_erase(&stable_node->node, root); 1255 struct rb_root *root; local 1262 root = root_stable_tree + nid; 1263 new = &root->rb_node; 1302 rb_insert_color(&stable_node->node, root); 1327 struct rb_root *root; local [all...] |
H A D | memory.c | 28 * 20.12.91 - Ok, making the swap-device changeable like the root. 2314 static inline void unmap_mapping_range_tree(struct rb_root *root, argument 2320 vma_interval_tree_foreach(vma, root,
|
H A D | mmap.c | 200 * Reserve some for root 213 * Reserve some for root 380 static int browse_rb(struct rb_root *root) argument 386 for (nd = rb_first(root); nd; nd = rb_next(nd)) { 425 static void validate_mm_rb(struct rb_root *root, struct vm_area_struct *ignore) argument 429 for (nd = rb_first(root); nd; nd = rb_next(nd)) { 474 #define validate_mm_rb(root, ignore) do { } while (0) 496 struct rb_root *root) 499 validate_mm_rb(root, NULL); 501 rb_insert_augmented(&vma->vm_rb, root, 495 vma_rb_insert(struct vm_area_struct *vma, struct rb_root *root) argument 504 vma_rb_erase(struct vm_area_struct *vma, struct rb_root *root) argument 735 struct rb_root *root = NULL; local [all...] |
H A D | shmem.c | 2889 static int shmem_show_options(struct seq_file *seq, struct dentry *root) argument 2891 struct shmem_sb_info *sbinfo = SHMEM_SB(root->d_sb);
|
H A D | vmscan.c | 2310 struct mem_cgroup *root = sc->target_mem_cgroup; local 2320 memcg = mem_cgroup_iter(root, NULL, &reclaim); 2342 mem_cgroup_iter_break(root, memcg); 2345 memcg = mem_cgroup_iter(root, memcg, &reclaim);
|
H A D | slab.c | 3737 struct kmem_cache *root = memcg_root_cache(cachep); local 3738 limit = root->limit; 3739 shared = root->shared; 3740 batchcount = root->batchcount;
|
H A D | memcontrol.c | 149 * matches memcg->dead_count of the hierarchy root group. 545 * process context. So the test for root memcg given 1067 * ref. count) or NULL if the whole root's subtree has been visited. 1071 static struct mem_cgroup *__mem_cgroup_iter_next(struct mem_cgroup *root, argument 1078 next_css = css_next_descendant_pre(prev_css, &root->css); 1087 * We do not take a reference on the root of the tree walk 1088 * because we might race with the root removal when it would 1098 if (next_css == &root->css) 1119 static void mem_cgroup_iter_invalidate(struct mem_cgroup *root) argument 1122 * When a group in the hierarchy below root i 1130 mem_cgroup_iter_load(struct mem_cgroup_reclaim_iter *iter, struct mem_cgroup *root, int *sequence) argument 1161 mem_cgroup_iter_update(struct mem_cgroup_reclaim_iter *iter, struct mem_cgroup *last_visited, struct mem_cgroup *new_position, struct mem_cgroup *root, int sequence) argument 1198 mem_cgroup_iter(struct mem_cgroup *root, struct mem_cgroup *prev, struct mem_cgroup_reclaim_cookie *reclaim) argument 1267 mem_cgroup_iter_break(struct mem_cgroup *root, struct mem_cgroup *prev) argument [all...] |