/mm/ |
H A D | fadvise.c | 27 SYSCALL_DEFINE(fadvise64_64)(int fd, loff_t offset, loff_t len, int advice) argument 47 if (!mapping || len < 0) { 69 endbyte = offset + len; 70 if (!len || endbyte < len) 138 asmlinkage long SyS_fadvise64_64(long fd, loff_t offset, loff_t len, long advice) argument 140 return SYSC_fadvise64_64((int) fd, offset, len, (int) advice); 147 SYSCALL_DEFINE(fadvise64)(int fd, loff_t offset, size_t len, int advice) argument 149 return sys_fadvise64_64(fd, offset, len, advice); 152 asmlinkage long SyS_fadvise64(long fd, loff_t offset, long len, lon argument [all...] |
H A D | util.c | 21 size_t len; local 27 len = strlen(s) + 1; 28 buf = kmalloc_track_caller(len, gfp); 30 memcpy(buf, s, len); 43 size_t len; local 49 len = strnlen(s, max); 50 buf = kmalloc_track_caller(len+1, gfp); 52 memcpy(buf, s, len); 53 buf[len] = '\0'; 63 * @len 66 kmemdup(const void *src, size_t len, gfp_t gfp) argument 85 memdup_user(const void __user *src, size_t len) argument [all...] |
H A D | msync.c | 31 SYSCALL_DEFINE3(msync, unsigned long, start, size_t, len, int, flags) 46 len = (len + ~PAGE_MASK) & PAGE_MASK; 47 end = start + len;
|
H A D | mlock.c | 367 static int do_mlock(unsigned long start, size_t len, int on) argument 374 VM_BUG_ON(len != PAGE_ALIGN(len)); 375 end = start + len; 417 static int do_mlock_pages(unsigned long start, size_t len, int ignore_errors) argument 426 VM_BUG_ON(len != PAGE_ALIGN(len)); 427 end = start + len; 473 SYSCALL_DEFINE2(mlock, unsigned long, start, size_t, len) 485 len [all...] |
H A D | mmap.c | 42 #define arch_mmap_check(addr, len, flags) (0) 46 #define arch_rebalance_pgtables(addr, len) (addr) 951 unsigned long len, unsigned long prot, 970 if (!len) 977 len = PAGE_ALIGN(len); 978 if (!len) 982 if ((pgoff + (len >> PAGE_SHIFT)) < pgoff) 992 addr = get_unmapped_area(file, addr, len, pgoff, flags); 1010 locked = len >> PAGE_SHIF 950 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff) argument 1132 unsigned long len; member in struct:mmap_arg_struct 1201 mmap_region(struct file *file, unsigned long addr, unsigned long len, unsigned long flags, vm_flags_t vm_flags, unsigned long pgoff) argument 1372 arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 1446 arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, const unsigned long len, const unsigned long pgoff, const unsigned long flags) argument 1545 get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 2028 do_munmap(struct mm_struct *mm, unsigned long start, size_t len) argument 2140 do_brk(unsigned long addr, unsigned long len) argument 2324 copy_vma(struct vm_area_struct **vmap, unsigned long addr, unsigned long len, pgoff_t pgoff) argument 2449 install_special_mapping(struct mm_struct *mm, unsigned long addr, unsigned long len, unsigned long vm_flags, struct page **pages) argument [all...] |
H A D | nommu.c | 854 unsigned long len) 857 unsigned long end = addr + len; 886 unsigned long len, 908 if (!len) 912 rlen = PAGE_ALIGN(len); 1125 unsigned long len, 1158 order = get_order(len); 1159 kdebug("alloc order %d for %lx", order, len); 1168 point = len >> PAGE_SHIFT; 1190 region->vm_end = region->vm_start + len; 852 find_vma_exact(struct mm_struct *mm, unsigned long addr, unsigned long len) argument 884 validate_mmap_request(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *_capabilities) argument 1123 do_mmap_private(struct vm_area_struct *vma, struct vm_region *region, unsigned long len, unsigned long capabilities) argument 1237 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff) argument 1505 unsigned long len; member in struct:mmap_arg_struct 1637 do_munmap(struct mm_struct *mm, unsigned long start, size_t len) argument 1748 do_brk(unsigned long addr, unsigned long len) argument 1844 arch_get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags) argument 1973 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 2015 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 2025 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | cleancache.c | 79 int len = 0, maxlen = CLEANCACHE_KEY_MAX; local 88 len = (*fhfn)(&d, &key->u.fh[0], &maxlen, 0); 89 if (len <= 0 || len == 255)
|
H A D | filemap_xip.c | 56 size_t len, 93 if (nr > len - copied) 94 nr = len - copied; 136 } while (copied < len); 147 xip_file_read(struct file *filp, char __user *buf, size_t len, loff_t *ppos) argument 149 if (!access_ok(VERIFY_WRITE, buf, len)) 153 buf, len, ppos); 395 xip_file_write(struct file *filp, const char __user *buf, size_t len, argument 406 if (!access_ok(VERIFY_READ, buf, len)) { 412 count = len; 52 do_xip_mapping_read(struct address_space *mapping, struct file_ra_state *_ra, struct file *filp, char __user *buf, size_t len, loff_t *ppos) argument [all...] |
H A D | ashmem.c | 215 size_t len, loff_t *pos) 232 ret = asma->file->f_op->read(asma->file, buf, len, pos); 436 size_t len; local 439 * Copying only `len', instead of ASHMEM_NAME_LEN, bytes 442 len = strlen(asma->name + ASHMEM_NAME_PREFIX_LEN) + 1; 444 asma->name + ASHMEM_NAME_PREFIX_LEN, len))) 594 /* per custom, you can pass zero for len to mean "everything onward" */ 595 if (!pin.len) 596 pin.len = PAGE_ALIGN(asma->size) - pin.offset; 598 if (unlikely((pin.offset | pin.len) 214 ashmem_read(struct file *file, char __user *buf, size_t len, loff_t *pos) argument [all...] |
H A D | mincore.c | 243 * current process's address space specified by [addr, addr + len). 257 * -ENOMEM - Addresses in the range [addr, addr + len] are 263 SYSCALL_DEFINE3(mincore, unsigned long, start, size_t, len, 275 if (!access_ok(VERIFY_READ, (void __user *) start, len)) 279 pages = len >> PAGE_SHIFT; 280 pages += (len & ~PAGE_MASK) != 0;
|
H A D | mprotect.c | 232 SYSCALL_DEFINE3(mprotect, unsigned long, start, size_t, len, 245 if (!len) 247 len = PAGE_ALIGN(len); 248 end = start + len;
|
H A D | madvise.c | 336 * -EINVAL - start + len < 0, start is not page-aligned, 352 size_t len; local 369 len = (len_in + ~PAGE_MASK) & PAGE_MASK; 371 /* Check to see whether len was rounded up from small -ve to zero */ 372 if (len_in && !len) 375 end = start + len;
|
H A D | slub.c | 3803 int len = 0; local 3837 if (len > PAGE_SIZE - KSYM_SYMBOL_LEN - 100) 3839 len += sprintf(buf + len, "%7ld ", l->count); 3842 len += sprintf(buf + len, "%pS", (void *)l->addr); 3844 len += sprintf(buf + len, "<not-available>"); 3847 len += sprintf(buf + len, " ag 4430 int len; local 4584 slab_attr_store(struct kobject *kobj, struct attribute *attr, const char *buf, size_t len) argument [all...] |
H A D | shmem.c | 1687 loff_t pos, unsigned len, unsigned flags, 1698 loff_t pos, unsigned len, unsigned copied, 2007 int len; local 2013 len = strlen(symname) + 1; 2014 if (len > PAGE_CACHE_SIZE) 2032 inode->i_size = len-1; 2033 if (len <= SHMEM_SYMLINK_INLINE_LEN) { 2035 memcpy(info->inline_symlink, symname, len); 2046 memcpy(kaddr, symname, len); 2127 size_t len; local 1686 shmem_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned flags, struct page **pagep, void **fsdata) argument 1697 shmem_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument 2191 struct { const char *prefix; size_t len; } arr[] = { member in struct:__anon10 2287 size_t len; local 2367 shmem_encode_fh(struct dentry *dentry, __u32 *fh, int *len, int connectable) argument [all...] |
H A D | mremap.c | 133 unsigned long new_addr, unsigned long len) 138 old_end = old_addr + len; 162 return len + old_addr - old_end; /* how much done */ 427 * We allow a zero old-len as a special case 429 * a zero new-len is nonsensical. 131 move_page_tables(struct vm_area_struct *vma, unsigned long old_addr, struct vm_area_struct *new_vma, unsigned long new_addr, unsigned long len) argument
|
H A D | mempolicy.c | 1077 static long do_mbind(unsigned long start, unsigned long len, argument 1100 len = (len + PAGE_SIZE - 1) & PAGE_MASK; 1101 end = start + len; 1120 start, start + len, mode, mode_flags, 1240 SYSCALL_DEFINE6(mbind, unsigned long, start, unsigned long, len, 1258 return do_mbind(start, len, mode, mode_flags, &nodes, flags); 1446 asmlinkage long compat_sys_mbind(compat_ulong_t start, compat_ulong_t len, argument 1467 return sys_mbind(start, len, mode, nm, nr_bits+1, flags);
|
H A D | memory.c | 3568 int ret, len, write; local 3582 len = DIV_ROUND_UP(end, PAGE_SIZE) - addr/PAGE_SIZE; 3584 len, write, 0, NULL, NULL); 3587 return ret == len ? 0 : -EFAULT; 3742 void *buf, int len, int write) 3754 memcpy_toio(maddr + offset, buf, len); 3756 memcpy_fromio(buf, maddr + offset, len); 3759 return len; 3768 unsigned long addr, void *buf, int len, int write) 3775 while (len) { 3741 generic_access_phys(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write) argument 3767 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3835 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3846 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | kmemleak.c | 287 int i, len, remaining; local 291 remaining = len = 294 seq_printf(seq, " hex dump (first %d bytes):\n", len); 295 for (i = 0; i < len; i += HEX_ROW_SIZE) {
|
H A D | filemap.c | 1491 unsigned long len = end - start + 1; local 1492 ret = do_readahead(mapping, file, start, len); 2217 loff_t pos, unsigned len, unsigned flags, 2222 return aops->write_begin(file, mapping, pos, len, flags, 2228 loff_t pos, unsigned len, unsigned copied, 2234 return aops->write_end(file, mapping, pos, len, copied, page, fsdata); 2216 pagecache_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned flags, struct page **pagep, void **fsdata) argument 2227 pagecache_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument
|
H A D | bootmem.c | 443 int __weak __init reserve_bootmem_generic(unsigned long phys, unsigned long len, argument 446 return reserve_bootmem(phys, len, flags);
|
H A D | hugetlb.c | 1378 const char *buf, size_t len) 1420 return len; 1433 struct kobj_attribute *attr, const char *buf, size_t len) 1435 return nr_hugepages_store_common(false, kobj, attr, buf, len); 1452 struct kobj_attribute *attr, const char *buf, size_t len) 1454 return nr_hugepages_store_common(true, kobj, attr, buf, len); 1376 nr_hugepages_store_common(bool obey_mempolicy, struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t len) argument 1432 nr_hugepages_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t len) argument 1451 nr_hugepages_mempolicy_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t len) argument
|
H A D | swapfile.c | 1756 int len; local 1764 len = seq_path(swap, &file->f_path, " \t\n\\"); 1766 len < 40 ? 40 - len : 1, " ",
|
H A D | slab.c | 298 static void free_block(struct kmem_cache *cachep, void **objpp, int len, 1889 "Slab corruption: %s start=%p, len=%d\n", 1917 printk(KERN_ERR "Prev obj: start=%p, len=%d\n", 1924 printk(KERN_ERR "Next obj: start=%p, len=%d\n",
|