/mm/ |
H A D | filemap_xip.c | 62 loff_t isize, pos; local 67 pos = *ppos; 68 index = pos >> PAGE_CACHE_SHIFT; 69 offset = pos & ~PAGE_CACHE_MASK; 119 * "pos" here (the actor routine has to update the user buffer 139 *ppos = pos + copied; 322 size_t count, loff_t pos, loff_t *ppos) 340 offset = (pos & (PAGE_CACHE_SIZE -1)); /* Within page */ 341 index = pos >> PAGE_CACHE_SHIFT; 371 pos 321 __xip_file_write(struct file *filp, const char __user *buf, size_t count, loff_t pos, loff_t *ppos) argument 401 loff_t pos; local [all...] |
H A D | ashmem.c | 215 size_t len, loff_t *pos) 232 ret = asma->file->f_op->read(asma->file, buf, len, pos); 237 /** Update backing file pos, since f_ops->read() doesn't */ 238 asma->file->f_pos = *pos; 214 ashmem_read(struct file *file, char __user *buf, size_t len, loff_t *pos) argument
|
H A D | page_cgroup.c | 423 unsigned long pos = offset & SC_POS_MASK; local 434 sc += pos; 458 unsigned long pos = offset & SC_POS_MASK; local 469 sc += pos; 489 unsigned long pos = offset & SC_POS_MASK; local 498 sc += pos;
|
H A D | bootmem.c | 327 unsigned long pos; local 330 pos = start; 335 if (pos < bdata->node_min_pfn || 336 pos >= bdata->node_low_pfn) { 337 BUG_ON(pos != start); 343 err = mark_bootmem_node(bdata, pos, max, reserve, flags); 345 mark_bootmem(start, pos, 0, 0); 351 pos = bdata->node_low_pfn;
|
H A D | vmstat.c | 619 static void *frag_start(struct seq_file *m, loff_t *pos) argument 622 loff_t node = *pos; 631 static void *frag_next(struct seq_file *m, void *arg, loff_t *pos) argument 635 (*pos)++; 1063 static void *vmstat_start(struct seq_file *m, loff_t *pos) argument 1068 if (*pos >= ARRAY_SIZE(vmstat_text)) 1094 return (unsigned long *)m->private + *pos; 1097 static void *vmstat_next(struct seq_file *m, void *arg, loff_t *pos) argument 1099 (*pos)++; 1100 if (*pos > [all...] |
H A D | filemap.c | 1174 * "pos" here (the actor routine has to update the user buffer 1368 * @pos: current file position 1375 unsigned long nr_segs, loff_t pos) 1399 if (pos < size) { 1400 retval = filemap_write_and_wait_range(mapping, pos, 1401 pos + iov_length(iov, nr_segs) - 1); 1407 iov, pos, nr_segs); 1411 *ppos = pos + retval; 2141 inline int generic_write_checks(struct file *file, loff_t *pos, size_t *count, int isblk) argument 2146 if (unlikely(*pos < 1374 generic_file_aio_read(struct kiocb *iocb, const struct iovec *iov, unsigned long nr_segs, loff_t pos) argument 2216 pagecache_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned flags, struct page **pagep, void **fsdata) argument 2227 pagecache_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument 2239 generic_file_direct_write(struct kiocb *iocb, const struct iovec *iov, unsigned long *nr_segs, loff_t pos, loff_t *ppos, size_t count, size_t ocount) argument 2342 generic_perform_write(struct file *file, struct iov_iter *i, loff_t pos) argument 2432 generic_file_buffered_write(struct kiocb *iocb, const struct iovec *iov, unsigned long nr_segs, loff_t pos, loff_t *ppos, size_t count, ssize_t written) argument 2479 loff_t pos; local 2579 generic_file_aio_write(struct kiocb *iocb, const struct iovec *iov, unsigned long nr_segs, loff_t pos) argument [all...] |
H A D | kmemleak.c | 1380 * a memory scanning when the pos argument points to the first position. 1382 static void *kmemleak_seq_start(struct seq_file *seq, loff_t *pos) argument 1385 loff_t n = *pos; 1408 static void *kmemleak_seq_next(struct seq_file *seq, void *v, loff_t *pos) argument 1414 ++(*pos);
|
H A D | percpu.c | 929 struct pcpu_chunk *pos; local 931 list_for_each_entry(pos, &pcpu_slot[pcpu_nr_slots - 1], list) 932 if (pos != chunk) {
|
H A D | swapfile.c | 1702 static void *swap_start(struct seq_file *swap, loff_t *pos) argument 1706 loff_t l = *pos; 1725 static void *swap_next(struct seq_file *swap, void *v, loff_t *pos) argument 1740 ++*pos;
|
H A D | vmalloc.c | 2492 static void *s_start(struct seq_file *m, loff_t *pos) 2495 loff_t n = *pos; 2511 static void *s_next(struct seq_file *m, void *p, loff_t *pos) argument 2515 ++*pos;
|
H A D | shmem.c | 1687 loff_t pos, unsigned len, unsigned flags, 1691 pgoff_t index = pos >> PAGE_CACHE_SHIFT; 1698 loff_t pos, unsigned len, unsigned copied, 1703 if (pos + copied > inode->i_size) 1704 i_size_write(inode, pos + copied); 1796 * "pos" here (the actor routine has to update the user buffer 1816 const struct iovec *iov, unsigned long nr_segs, loff_t pos) 1686 shmem_write_begin(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned flags, struct page **pagep, void **fsdata) argument 1697 shmem_write_end(struct file *file, struct address_space *mapping, loff_t pos, unsigned len, unsigned copied, struct page *page, void *fsdata) argument 1815 shmem_file_aio_read(struct kiocb *iocb, const struct iovec *iov, unsigned long nr_segs, loff_t pos) argument
|
H A D | slab.c | 4170 static void *s_start(struct seq_file *m, loff_t *pos) argument 4172 loff_t n = *pos; 4178 return seq_list_start(&cache_chain, *pos); 4181 static void *s_next(struct seq_file *m, void *p, loff_t *pos) argument 4183 return seq_list_next(p, &cache_chain, pos); 4375 static void *leaks_start(struct seq_file *m, loff_t *pos) argument 4378 return seq_list_start(&cache_chain, *pos);
|
H A D | slub.c | 3712 long start, end, pos; local 3721 pos = start + (end - start + 1) / 2; 3727 if (pos == end) 3730 caddr = t->loc[pos].addr; 3733 l = &t->loc[pos]; 3755 end = pos; 3757 start = pos; 3766 l = t->loc + pos; 3767 if (pos < t->count) 3769 (t->count - pos) * sizeo 4826 s_start(struct seq_file *m, loff_t *pos) argument 4837 s_next(struct seq_file *m, void *p, loff_t *pos) argument [all...] |
H A D | page_alloc.c | 2962 int pos, j, node; local 2968 pos = 0; 2975 &zonelist->_zonerefs[pos++]); 2980 zonelist->_zonerefs[pos].zone = NULL; 2981 zonelist->_zonerefs[pos].zone_idx = 0;
|