Lines Matching refs:write
1666 * pages. Otherwise, someone (memory error handling) may try to write
2163 * Find and lock address space (mapping) in write mode.
4919 static int proc_hugetlb_doulongvec_minmax(struct ctl_table *table, int write,
4932 return proc_doulongvec_minmax(&dup_table, write, buffer, length, ppos);
4936 struct ctl_table *table, int write,
4946 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos,
4951 if (write)
4958 static int hugetlb_sysctl_handler(struct ctl_table *table, int write,
4962 return hugetlb_sysctl_handler_common(false, table, write,
4967 static int hugetlb_mempolicy_sysctl_handler(struct ctl_table *table, int write,
4970 return hugetlb_sysctl_handler_common(true, table, write,
4975 static int hugetlb_overcommit_handler(struct ctl_table *table, int write,
4987 if (write && hstate_is_gigantic(h))
4990 ret = proc_hugetlb_doulongvec_minmax(table, write, buffer, length, ppos,
4995 if (write) {
5953 * hugetlb does not support FOLL_FORCE-style write faults that keep the
6027 * in write mode. Dropping vma_lock in read mode
6278 * does a clear then write of pte's under page table
6775 * set_pte_at() write. The memory barrier inside __folio_mark_uptodate
6781 * before the set_pte_at() write.
6829 * with wp flag set, don't set pte write bit.
6841 * thrown away at will, even if write bit not set.
7131 * called to make the mapping read-write. Assume !vma is a shm mapping