/mm/ |
H A D | fadvise.c | 27 SYSCALL_DEFINE(fadvise64_64)(int fd, loff_t offset, loff_t len, int advice) argument 47 if (!mapping || len < 0) { 69 endbyte = offset + len; 70 if (!len || endbyte < len) 139 asmlinkage long SyS_fadvise64_64(long fd, loff_t offset, loff_t len, long advice) argument 141 return SYSC_fadvise64_64((int) fd, offset, len, (int) advice); 148 SYSCALL_DEFINE(fadvise64)(int fd, loff_t offset, size_t len, int advice) argument 150 return sys_fadvise64_64(fd, offset, len, advice); 153 asmlinkage long SyS_fadvise64(long fd, loff_t offset, long len, lon argument [all...] |
H A D | cleancache.c | 79 int len = 0, maxlen = CLEANCACHE_KEY_MAX; local 88 len = (*fhfn)(&d, &key->u.fh[0], &maxlen, 0); 89 if (len <= 0 || len == 255)
|
H A D | madvise.c | 336 * -EINVAL - start + len < 0, start is not page-aligned, 352 size_t len; local 369 len = (len_in + ~PAGE_MASK) & PAGE_MASK; 371 /* Check to see whether len was rounded up from small -ve to zero */ 372 if (len_in && !len) 375 end = start + len;
|
H A D | filemap_xip.c | 56 size_t len, 93 if (nr > len - copied) 94 nr = len - copied; 136 } while (copied < len); 147 xip_file_read(struct file *filp, char __user *buf, size_t len, loff_t *ppos) argument 149 if (!access_ok(VERIFY_WRITE, buf, len)) 153 buf, len, ppos); 395 xip_file_write(struct file *filp, const char __user *buf, size_t len, argument 406 if (!access_ok(VERIFY_READ, buf, len)) { 412 count = len; 52 do_xip_mapping_read(struct address_space *mapping, struct file_ra_state *_ra, struct file *filp, char __user *buf, size_t len, loff_t *ppos) argument [all...] |
H A D | mlock.c | 375 static int do_mlock(unsigned long start, size_t len, int on) argument 382 VM_BUG_ON(len != PAGE_ALIGN(len)); 383 end = start + len; 425 static int do_mlock_pages(unsigned long start, size_t len, int ignore_errors) argument 434 VM_BUG_ON(len != PAGE_ALIGN(len)); 435 end = start + len; 481 SYSCALL_DEFINE2(mlock, unsigned long, start, size_t, len) 493 len [all...] |
H A D | mremap.c | 125 unsigned long new_addr, unsigned long len) 131 old_end = old_addr + len; 176 flush_tlb_range(vma, old_end-len, old_addr); 178 mmu_notifier_invalidate_range_end(vma->vm_mm, old_end-len, old_end); 180 return len + old_addr - old_end; /* how much done */ 454 * We allow a zero old-len as a special case 456 * a zero new-len is nonsensical. 123 move_page_tables(struct vm_area_struct *vma, unsigned long old_addr, struct vm_area_struct *new_vma, unsigned long new_addr, unsigned long len) argument
|
H A D | process_vm_access.c | 32 * @len: number of bytes to copy 47 unsigned long len, 95 len - *bytes_copied); 157 * @len: size of area to copy to/from 171 unsigned long len, 195 if (len == 0) 197 nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1; 204 start_offset, len, 215 len -= bytes_copied_loop; 42 process_vm_rw_pages(struct task_struct *task, struct mm_struct *mm, struct page **process_pages, unsigned long pa, unsigned long start_offset, unsigned long len, const struct iovec *lvec, unsigned long lvec_cnt, unsigned long *lvec_current, size_t *lvec_offset, int vm_write, unsigned int nr_pages_to_copy, ssize_t *bytes_copied) argument 170 process_vm_rw_single_vec(unsigned long addr, unsigned long len, const struct iovec *lvec, unsigned long lvec_cnt, unsigned long *lvec_current, size_t *lvec_offset, struct page **process_pages, struct mm_struct *mm, struct task_struct *task, int vm_write, ssize_t *bytes_copied) argument
|
H A D | util.c | 21 size_t len; local 27 len = strlen(s) + 1; 28 buf = kmalloc_track_caller(len, gfp); 30 memcpy(buf, s, len); 43 size_t len; local 49 len = strnlen(s, max); 50 buf = kmalloc_track_caller(len+1, gfp); 52 memcpy(buf, s, len); 53 buf[len] = '\0'; 63 * @len 66 kmemdup(const void *src, size_t len, gfp_t gfp) argument 85 memdup_user(const void __user *src, size_t len) argument [all...] |
H A D | bootmem.c | 441 int __weak __init reserve_bootmem_generic(unsigned long phys, unsigned long len, argument 444 return reserve_bootmem(phys, len, flags);
|
H A D | filemap.c | 1512 unsigned long len = end - start + 1; local 1513 ret = do_readahead(mapping, file, start, len); 2241 loff_t pos, unsigned len, unsigned flags, 2246 return aops->write_begin(file, mapping, pos, len, flags, 2252 loff_t pos, unsigned len, unsigned copied, 2258 return aops->write_end(file, mapping, pos, len, copied, page, fsdata); 2240 pagecache_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned flags, struct page **pagep, void **fsdata) argument 2251 pagecache_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument
|
H A D | kmemleak.c | 293 int i, len, remaining; local 297 remaining = len = 300 seq_printf(seq, " hex dump (first %d bytes):\n", len); 301 for (i = 0; i < len; i += HEX_ROW_SIZE) {
|
H A D | memory.c | 3580 int ret, len, write; local 3594 len = DIV_ROUND_UP(end, PAGE_SIZE) - addr/PAGE_SIZE; 3596 len, write, 0, NULL, NULL); 3599 return ret == len ? 0 : -EFAULT; 3754 void *buf, int len, int write) 3766 memcpy_toio(maddr + offset, buf, len); 3768 memcpy_fromio(buf, maddr + offset, len); 3771 return len; 3780 unsigned long addr, void *buf, int len, int write) 3787 while (len) { 3753 generic_access_phys(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write) argument 3779 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3847 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3858 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | mempolicy.c | 1086 static long do_mbind(unsigned long start, unsigned long len, argument 1109 len = (len + PAGE_SIZE - 1) & PAGE_MASK; 1110 end = start + len; 1129 start, start + len, mode, mode_flags, 1249 SYSCALL_DEFINE6(mbind, unsigned long, start, unsigned long, len, 1267 return do_mbind(start, len, mode, mode_flags, &nodes, flags); 1457 asmlinkage long compat_sys_mbind(compat_ulong_t start, compat_ulong_t len, argument 1478 return sys_mbind(start, len, mode, nm, nr_bits+1, flags);
|
H A D | mmap.c | 42 #define arch_mmap_check(addr, len, flags) (0) 46 #define arch_rebalance_pgtables(addr, len) (addr) 943 unsigned long len, unsigned long prot, 962 if (!len) 969 len = PAGE_ALIGN(len); 970 if (!len) 974 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) 984 addr = get_unmapped_area(file, addr, len, pgoff, flags); 1002 locked = len >> PAGE_SHIF 942 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff) argument 1124 unsigned long len; member in struct:mmap_arg_struct 1193 mmap_region(struct file *file, unsigned long addr, unsigned long len, unsigned long flags, vm_flags_t vm_flags, unsigned long pgoff) argument 1364 arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 1438 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, const unsigned long len, const unsigned long pgoff, const unsigned long flags) argument 1537 get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 2000 do_munmap(struct mm_struct *mm, unsigned long start, size_t len) argument 2112 do_brk(unsigned long addr, unsigned long len) argument 2296 copy_vma(struct vm_area_struct **vmap, unsigned long addr, unsigned long len, pgoff_t pgoff) argument 2439 install_special_mapping(struct mm_struct *mm, unsigned long addr, unsigned long len, unsigned long vm_flags, struct page **pages) argument [all...] |
H A D | nommu.c | 853 unsigned long len) 856 unsigned long end = addr + len; 885 unsigned long len, 907 if (!len) 911 rlen = PAGE_ALIGN(len); 1124 unsigned long len, 1157 order = get_order(len); 1158 kdebug("alloc order %d for %lx", order, len); 1167 point = len >> PAGE_SHIFT; 1189 region->vm_end = region->vm_start + len; 851 find_vma_exact(struct mm_struct *mm, unsigned long addr, unsigned long len) argument 883 validate_mmap_request(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *_capabilities) argument 1122 do_mmap_private(struct vm_area_struct *vma, struct vm_region *region, unsigned long len, unsigned long capabilities) argument 1236 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff) argument 1504 unsigned long len; member in struct:mmap_arg_struct 1636 do_munmap(struct mm_struct *mm, unsigned long start, size_t len) argument 1747 do_brk(unsigned long addr, unsigned long len) argument 1843 arch_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 1964 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 2006 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 2016 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | shmem.c | 1183 loff_t pos, unsigned len, unsigned flags, 1193 loff_t pos, unsigned len, unsigned copied, 1347 struct pipe_inode_info *pipe, size_t len, 1372 if (unlikely(left < len)) 1373 len = left; 1380 req_pages = (len + loff + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; 1404 if (!len) 1407 this_len = min_t(unsigned long, len, PAGE_CACHE_SIZE - loff); 1433 len = this_len; 1437 spd.partial[page_nr].len 1182 shmem_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned flags, struct page **pagep, void **fsdata) argument 1192 shmem_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument 1346 shmem_file_splice_read(struct file *in, loff_t *ppos, struct pipe_inode_info *pipe, size_t len, unsigned int flags) argument 1618 int len; local 1741 size_t len; local 1804 struct { const char *prefix; size_t len; } arr[] = { member in struct:__anon11 1900 size_t len; local 1980 shmem_encode_fh(struct dentry *dentry, __u32 *fh, int *len, int connectable) argument [all...] |
H A D | swapfile.c | 1753 int len; local 1761 len = seq_path(swap, &file->f_path, " \t\n\\"); 1763 len < 40 ? 40 - len : 1, " ",
|
H A D | hugetlb.c | 1386 const char *buf, size_t len) 1428 return len; 1441 struct kobj_attribute *attr, const char *buf, size_t len) 1443 return nr_hugepages_store_common(false, kobj, attr, buf, len); 1460 struct kobj_attribute *attr, const char *buf, size_t len) 1462 return nr_hugepages_store_common(true, kobj, attr, buf, len); 1384 nr_hugepages_store_common(bool obey_mempolicy, struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t len) argument 1440 nr_hugepages_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t len) argument 1459 nr_hugepages_mempolicy_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t len) argument
|
H A D | slub.c | 4289 int len = 0; local 4323 if (len > PAGE_SIZE - KSYM_SYMBOL_LEN - 100) 4325 len += sprintf(buf + len, "%7ld ", l->count); 4328 len += sprintf(buf + len, "%pS", (void *)l->addr); 4330 len += sprintf(buf + len, "<not-available>"); 4333 len += sprintf(buf + len, " ag 4712 int len; local 4990 int len; local 5158 slab_attr_store(struct kobject *kobj, struct attribute *attr, const char *buf, size_t len) argument [all...] |