/mm/ |
H A D | fadvise.c | 37 int ret = 0; local 43 ret = -ESPIPE; 49 ret = -EINVAL; 64 ret = -EINVAL; 142 ret = -EINVAL; 146 return ret;
|
H A D | maccess.c | 23 long ret; local 28 ret = __copy_from_user_inatomic(dst, 33 return ret ? -EFAULT : 0; 51 long ret; local 56 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); 60 return ret ? -EFAULT : 0;
|
H A D | list_lru.c | 78 enum lru_status ret; local 88 ret = isolate(item, &nlru->lock, cb_arg); 89 switch (ret) { 102 if (ret == LRU_REMOVED_RETRY)
|
H A D | mmu_notifier.c | 208 int ret; local 218 ret = -ENOMEM; 225 ret = mm_take_all_locks(mm); 226 if (unlikely(ret)) 257 return ret;
|
H A D | page_io.c | 150 int ret; local 204 ret = add_swap_extent(sis, page_no, 1, first_block); 205 if (ret < 0) 207 nr_extents += ret; 213 ret = nr_extents; 221 return ret; 224 ret = -EINVAL; 234 int ret = 0; local 246 ret = __swap_writepage(page, wbc, end_swap_bio_write); 248 return ret; 260 int ret, rw = WRITE; local 339 int ret = 0; local [all...] |
H A D | page_isolation.c | 18 int ret = -EBUSY; local 50 ret = 0; 58 if (!ret) { 70 if (!ret) 72 return ret; 266 int ret; local 284 ret = __test_page_isolated_in_pageblock(start_pfn, end_pfn, 287 return ret ? 0 : -EBUSY;
|
H A D | workingset.c | 311 int ret; local 331 ret = LRU_RETRY; 363 ret = LRU_REMOVED_RETRY; 369 return ret; 375 unsigned long ret; local 379 ret = list_lru_walk_node(&workingset_shadow_nodes, sc->nid, 382 return ret; 400 int ret; local 402 ret = list_lru_init_key(&workingset_shadow_nodes, &shadow_nodes_key); 403 if (ret) [all...] |
H A D | zpool.c | 61 int ret = 0, refcount; local 67 ret = -EBUSY; 72 return ret;
|
H A D | cleancache.c | 230 int ret = -1; local 250 ret = cleancache_ops->get_page(pool_id, 252 if (ret == 0) 257 return ret;
|
H A D | cma.c | 136 int ret = cma_activate_area(&cma_areas[i]); local 138 if (ret) 139 return ret; 219 int ret = 0; local 259 ret = -EINVAL; 277 ret = -EBUSY; 299 ret = -ENOMEM; 307 ret = cma_init_reserved_mem(base, size, order_per_bit, res_cma); 308 if (ret) 317 return ret; 334 int ret; local [all...] |
H A D | filemap_xip.c | 275 int err, ret = VM_FAULT_OOM; local 297 ret = VM_FAULT_NOPAGE; 302 return ret; 405 ssize_t ret; local 410 ret=-EFAULT; 420 ret = generic_write_checks(filp, &pos, &count, S_ISBLK(inode->i_mode)); 421 if (ret) 426 ret = file_remove_suid(filp); 427 if (ret) 430 ret [all...] |
H A D | frontswap.c | 194 bool ret = false; local 197 ret = test_bit(offset, sis->frontswap_map); 198 return ret; 218 int ret = -1, dup = 0; local 229 return ret; 235 ret = frontswap_ops->store(type, offset, page); 236 if (ret == 0) { 254 ret = -1; 255 return ret; 266 int ret local 344 int ret = -EINVAL; local 408 int uninitialized_var(type), ret; local [all...] |
H A D | gup.c | 229 int ret = -EFAULT; local 259 ret = 0; 262 return ret; 275 int ret; local 293 ret = handle_mm_fault(mm, vma, address, fault_flags); 294 if (ret & VM_FAULT_ERROR) { 295 if (ret & VM_FAULT_OOM) 297 if (ret & (VM_FAULT_HWPOISON | VM_FAULT_HWPOISON_LARGE)) 299 if (ret & VM_FAULT_SIGBUS) 305 if (ret 456 int ret; local 485 int ret; local 558 int ret; local 727 int ret = 0; local 1004 int nr, ret; local [all...] |
H A D | highmem.c | 415 void *ret; local 422 ret = NULL; 429 ret = pam->virtual; 436 return ret;
|
H A D | hugetlb_cgroup.c | 168 int ret = 0; local 190 ret = res_counter_charge(&h_cg->hugepage[idx], csize, &fail_res); 194 return ret; 259 int idx, name, ret; local 271 ret = -EINVAL; 275 ret = res_counter_memparse_write_strategy(buf, &val); 276 if (ret) 279 ret = res_counter_set_limit(&h_cg->hugepage[idx], val); 282 ret = -EINVAL; 285 return ret 291 int idx, name, ret = 0; local [all...] |
H A D | madvise.c | 349 int ret; local 351 ret = get_user_pages_fast(start, 1, 0, &p); 352 if (ret != 1) 353 return ret; 362 ret = soft_offline_page(p, MF_COUNT_INCREASED); 363 if (ret) 364 return ret;
|
H A D | mremap.c | 400 unsigned long ret = -EINVAL; local 419 ret = do_munmap(mm, new_addr, new_len); 420 if (ret) 424 ret = do_munmap(mm, addr+new_len, old_len - new_len); 425 if (ret && old_len != new_len) 432 ret = PTR_ERR(vma); 440 ret = get_unmapped_area(vma->vm_file, new_addr, new_len, vma->vm_pgoff + 443 if (ret & ~PAGE_MASK) 446 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked); 447 if (!(ret 482 unsigned long ret = -EINVAL; local [all...] |
H A D | oom_kill.c | 61 bool ret = false; local 72 ret = mempolicy_nodemask_intersects(tsk, mask); 78 ret = cpuset_mems_allowed_intersects(current, tsk); 80 if (ret) 85 return ret; 581 bool ret = true; local 586 ret = false; 599 return ret;
|
H A D | readahead.c | 87 int ret = 0; local 99 ret = filler(data, page); 100 if (unlikely(ret)) { 106 return ret; 116 int ret; local 121 ret = mapping->a_ops->readpages(filp, mapping, pages, nr_pages); 136 ret = 0; 141 return ret; 161 int ret = 0; local 191 ret 564 ssize_t ret; local [all...] |
H A D | swap_state.c | 62 unsigned long ret = 0; local 65 ret += swapper_spaces[i].nrpages; 66 return ret;
|
H A D | truncate.c | 161 int ret; local 169 ret = remove_mapping(mapping, page); 171 return ret; 485 unsigned long ret; local 509 ret = invalidate_inode_page(page); 515 if (!ret) 517 count += ret; 589 int ret = 0; local 646 ret = ret2; 655 return ret; [all...] |
H A D | util.c | 243 int ret; local 246 ret = get_user_pages(current, mm, start, nr_pages, 250 return ret; 258 unsigned long ret; local 262 ret = security_mmap_file(file, prot, flag); 263 if (!ret) { 265 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, 269 mm_populate(ret, populate); 271 return ret; 318 int ret; local 330 int ret; local [all...] |
H A D | mlock.c | 122 int ret = SWAP_AGAIN; local 129 ret = try_to_munlock(page); 132 if (ret != SWAP_MLOCK) 560 int ret = 0; local 577 ret = split_vma(mm, vma, start, 1); 578 if (ret) 583 ret = split_vma(mm, vma, end, 0); 584 if (ret) 610 return ret; 677 long ret local 761 int ret; local 802 int ret = -EINVAL; local 832 int ret; local [all...] |
H A D | page_cgroup.c | 245 int ret = 0; local 248 ret = online_page_cgroup(mn->start_pfn, 266 return notifier_from_errno(ret);
|
H A D | slob.c | 431 void *ret; local 446 ret = (void *)m + align; 448 trace_kmalloc_node(caller, ret, 455 ret = slob_new_pages(gfp, order, node); 457 trace_kmalloc_node(caller, ret, 461 kmemleak_alloc(ret, size, 1, gfp); 462 return ret;
|