Commit cda540ac authored by Hugh Dickins's avatar Hugh Dickins Committed by Linus Torvalds
Browse files

mm: get_user_pages(write,force) refuse to COW in shared areas

get_user_pages(write=1, force=1) has always had odd behaviour on write-
protected shared mappings: although it demands FMODE_WRITE-access to the
underlying object (do_mmap_pgoff sets neither VM_SHARED nor VM_MAYWRITE
without that), it ends up with do_wp_page substituting private anonymous
Copied-On-Write pages for the shared file pages in the area.

That was long ago intentional, as a safety measure to prevent ptrace
setting a breakpoint (or POKETEXT or POKEDATA) from inadvertently
corrupting the underlying executable.  Yet exec and dynamic loaders open
the file read-only, and use MAP_PRIVATE rather than MAP_SHARED.

The traditional odd behaviour still causes surprises and bugs in mm, and
is probably not what any caller wants - even the comment on the flag
says "You do not want this" (although it's undoubtedly necessary for
overriding userspace protections in some contexts, and good when !write).

Let's stop doing that.  But it would be dangerous to remove the long-
standing safety at this stage, so just make get_user_pages(write,force)
fail with EFAULT when applied to a write-protected shared area.
Infiniband may in future want to force write through to underlying
object: we can add another FOLL_flag later to enable that if required.

Odd though the old behaviour was, there is no doubt that we may turn out
to break userspace with this change, and have to revert it quickly.
Issue a WARN_ON_ONCE to help debug the changed case (easily triggered by
userspace, so only once to prevent spamming the logs); and delay a few
associated cleanups until this change is proved.

get_user_pages callers who might see trouble from this change:
  ptrace poking, or writing to /proc/<pid>/mem
if they ever apply get_user_pages to write-protected shared mappings
of an object which was opened for writing.

I went to apply the same change to mm/nommu.c, but retreated.  NOMMU has
no place for COW, and its VM_flags conventions are not the same: I'd be
more likely to screw up NOMMU than make an improvement there.
Suggested-by: default avatarLinus Torvalds <>
Signed-off-by: default avatarHugh Dickins <>
Acked-by: default avatarKirill A. Shutemov <>
Signed-off-by: default avatarLinus Torvalds <>
parent d15e0310
......@@ -1705,15 +1705,6 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
VM_BUG_ON(!!pages != !!(gup_flags & FOLL_GET));
* Require read or write permissions.
* If FOLL_FORCE is set, we only require the "MAY" flags.
vm_flags = (gup_flags & FOLL_WRITE) ?
vm_flags &= (gup_flags & FOLL_FORCE) ?
* If FOLL_FORCE and FOLL_NUMA are both set, handle_mm_fault
* would be called on PROT_NONE ranges. We must never invoke
......@@ -1741,7 +1732,7 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
/* user gate pages are read-only */
if (gup_flags & FOLL_WRITE)
return i ? : -EFAULT;
goto efault;
if (pg > TASK_SIZE)
pgd = pgd_offset_k(pg);
......@@ -1751,12 +1742,12 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
pmd = pmd_offset(pud, pg);
if (pmd_none(*pmd))
return i ? : -EFAULT;
goto efault;
pte = pte_offset_map(pmd, pg);
if (pte_none(*pte)) {
return i ? : -EFAULT;
goto efault;
vma = get_gate_vma(mm);
if (pages) {
......@@ -1769,7 +1760,7 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
page = pte_page(*pte);
else {
return i ? : -EFAULT;
goto efault;
pages[i] = page;
......@@ -1780,10 +1771,42 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
goto next_page;
if (!vma ||
(vma->vm_flags & (VM_IO | VM_PFNMAP)) ||
!(vm_flags & vma->vm_flags))
return i ? : -EFAULT;
if (!vma)
goto efault;
vm_flags = vma->vm_flags;
if (vm_flags & (VM_IO | VM_PFNMAP))
goto efault;
if (gup_flags & FOLL_WRITE) {
if (!(vm_flags & VM_WRITE)) {
if (!(gup_flags & FOLL_FORCE))
goto efault;
* We used to let the write,force case do COW
* ptrace could set a breakpoint in a read-only
* mapping of an executable, without corrupting
* the file (yet only when that file had been
* opened for writing!). Anon pages in shared
* mappings are surprising: now just reject it.
if (!is_cow_mapping(vm_flags)) {
goto efault;
} else {
if (!(vm_flags & VM_READ)) {
if (!(gup_flags & FOLL_FORCE))
goto efault;
* Is there actually any vma we can reach here
* which does not have VM_MAYREAD set?
if (!(vm_flags & VM_MAYREAD))
goto efault;
if (is_vm_hugetlb_page(vma)) {
i = follow_hugetlb_page(mm, vma, pages, vmas,
......@@ -1837,7 +1860,7 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
return -EFAULT;
if (ret & VM_FAULT_SIGBUS)
return i ? i : -EFAULT;
goto efault;
......@@ -1895,6 +1918,8 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
} while (nr_pages && start < vma->vm_end);
} while (nr_pages);
return i;
return i ? : -EFAULT;
......@@ -1962,9 +1987,8 @@ int fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm,
* @start: starting user address
* @nr_pages: number of pages from start to pin
* @write: whether pages will be written to by the caller
* @force: whether to force write access even if user mapping is
* readonly. This will result in the page being COWed even
* in MAP_SHARED mappings. You do not want this.
* @force: whether to force access even when user mapping is currently
* protected (but never forces write access to shared mapping).
* @pages: array that receives pointers to the pages pinned.
* Should be at least nr_pages long. Or NULL, if caller
* only intends to ensure the pages are faulted in.
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment