/mm/ |
H A D | madvise.c | 111 * vm_flags is protected by the mmap_sem held in write mode. 161 * that was kept in them. There is no reason to write this data out to 351 int write; local 361 write = madvise_need_mmap_write(behavior); 362 if (write) 427 if (write)
|
H A D | util.c | 258 int nr_pages, int write, struct page **pages) 268 * @write: whether pages will be written to 289 int nr_pages, int write, struct page **pages) 296 write, 0, pages, NULL); 257 __get_user_pages_fast(unsigned long start, int nr_pages, int write, struct page **pages) argument 288 get_user_pages_fast(unsigned long start, int nr_pages, int write, struct page **pages) argument
|
H A D | compaction.c | 731 int sysctl_compaction_handler(struct ctl_table *table, int write, argument 734 if (write) 740 int sysctl_extfrag_handler(struct ctl_table *table, int write, argument 743 proc_dointvec_minmax(table, write, buffer, length, ppos);
|
H A D | memory.c | 905 * If it's a COW mapping, write protect it both 1614 * Must be called with mmap_sem held for read or write. 1657 * Require read or write permissions. 1797 * to write to the gotten user page, which a 1799 * page might get reCOWed by userspace write). 1890 * @write: whether pages will be written to by the caller 1891 * @force: whether to force write access even if user mapping is 1906 * Must be called with mmap_sem held for read or write. 1922 * If write=0, the page must not be written to. If the page is written to, 1935 unsigned long start, int nr_pages, int write, in 1934 get_user_pages(struct task_struct *tsk, struct mm_struct *mm, unsigned long start, int nr_pages, int write, int force, struct page **pages, struct vm_area_struct **vmas) argument 3580 int ret, len, write; local 3753 generic_access_phys(struct vm_area_struct *vma, unsigned long addr, void *buf, int len, int write) argument 3779 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3847 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 3858 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | nommu.c | 136 /* calculate required read or write permissions. 178 unsigned long start, int nr_pages, int write, int force, 183 if (write) 1965 unsigned long addr, void *buf, int len, int write) 1978 /* only read or write mappings where it is permitted */ 1979 if (write && vma->vm_flags & VM_MAYWRITE) 1982 else if (!write && vma->vm_flags & VM_MAYREAD) 2002 * @write: whether the access is a write 2007 void *buf, int len, int write) 177 get_user_pages(struct task_struct *tsk, struct mm_struct *mm, unsigned long start, int nr_pages, int write, int force, struct page **pages, struct vm_area_struct **vmas) argument 1964 __access_remote_vm(struct task_struct *tsk, struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 2006 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, int write) argument 2016 access_process_vm(struct task_struct *tsk, unsigned long addr, void *buf, int len, int write) argument [all...] |
H A D | page-writeback.c | 51 * Estimate write bandwidth at 200ms intervals. 124 * writeback completions [end_page_writeback()]. Those devices that write out 351 int dirty_background_ratio_handler(struct ctl_table *table, int write, argument 357 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); 358 if (ret == 0 && write) 363 int dirty_background_bytes_handler(struct ctl_table *table, int write, argument 369 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); 370 if (ret == 0 && write) 375 int dirty_ratio_handler(struct ctl_table *table, int write, argument 382 ret = proc_dointvec_minmax(table, write, buffe 390 dirty_bytes_handler(struct ctl_table *table, int write, void __user *buffer, size_t *lenp, loff_t *ppos) argument 1500 dirty_writeback_centisecs_handler(ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument [all...] |
H A D | hugetlb.c | 62 * must either hold the mmap_sem for write, or the mmap_sem for read and 1885 struct ctl_table *table, int write, 1894 if (write && h->order >= MAX_ORDER) 1899 ret = proc_doulongvec_minmax(table, write, buffer, length, ppos); 1903 if (write) { 1920 int hugetlb_sysctl_handler(struct ctl_table *table, int write, argument 1924 return hugetlb_sysctl_handler_common(false, table, write, 1929 int hugetlb_mempolicy_sysctl_handler(struct ctl_table *table, int write, argument 1932 return hugetlb_sysctl_handler_common(true, table, write, 1937 int hugetlb_treat_movable_handler(struct ctl_table *table, int write, argument 1884 hugetlb_sysctl_handler_common(bool obey_mempolicy, struct ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument 1949 hugetlb_overcommit_handler(struct ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument 2743 follow_huge_pud(struct mm_struct *mm, unsigned long address, pud_t *pud, int write) argument [all...] |
H A D | vmscan.c | 420 /* lumpy reclaim for hugepage often need a lot of write */ 427 * We detected a synchronous write error writing a page out. Probably 449 /* failed to write page out, page is locked */ 467 * If the page is dirty, only perform writeback if that write 477 * If the page is swapcache, write it back even if that would 523 /* synchronous write or broken a_ops? */ 902 /* Page is dirty, try to write it out here */ 916 * A synchronous write - probably a ramdisk. Go 1083 * is used by reclaim when it is cannot write to backing storage 2377 * Try to write bac 3572 scan_unevictable_handler(struct ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument [all...] |
H A D | page_alloc.c | 1738 * write pages from its LRU list. 2343 /* Wait for some write requests to complete then retry */ 2898 int numa_zonelist_order_handler(ctl_table *table, int write, argument 2907 if (write) 2909 ret = proc_dostring(table, write, buffer, length, ppos); 2912 if (write) { 5089 int min_free_kbytes_sysctl_handler(ctl_table *table, int write, 5092 proc_dointvec(table, write, buffer, length, ppos); 5093 if (write) 5099 int sysctl_min_unmapped_ratio_sysctl_handler(ctl_table *table, int write, argument 5115 sysctl_min_slab_ratio_sysctl_handler(ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument 5141 lowmem_reserve_ratio_sysctl_handler(ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument 5155 percpu_pagelist_fraction_sysctl_handler(ctl_table *table, int write, void __user *buffer, size_t *length, loff_t *ppos) argument [all...] |