Re: [PATCH v2] mm: mprotect: check page dirty when change ptes

From: Peter Xu
Date: Fri Sep 14 2018 - 03:16:22 EST


On Thu, Sep 13, 2018 at 08:42:39PM -0400, Jerome Glisse wrote:
> On Thu, Sep 13, 2018 at 10:23:28AM -0400, Jerome Glisse wrote:
> > On Thu, Sep 13, 2018 at 03:37:22PM +0800, Peter Xu wrote:
> > > On Wed, Sep 12, 2018 at 09:24:39AM -0400, Jerome Glisse wrote:
> > > > On Wed, Sep 12, 2018 at 09:03:55AM -0400, Jerome Glisse wrote:
> > > > > On Wed, Sep 12, 2018 at 02:49:21PM +0800, Peter Xu wrote:
> > > > > > Add an extra check on page dirty bit in change_pte_range() since there
> > > > > > might be case where PTE dirty bit is unset but it's actually dirtied.
> > > > > > One example is when a huge PMD is splitted after written: the dirty bit
> > > > > > will be set on the compound page however we won't have the dirty bit set
> > > > > > on each of the small page PTEs.
> > > > > >
> > > > > > I noticed this when debugging with a customized kernel that implemented
> > > > > > userfaultfd write-protect. In that case, the dirty bit will be critical
> > > > > > since that's required for userspace to handle the write protect page
> > > > > > fault (otherwise it'll get a SIGBUS with a loop of page faults).
> > > > > > However it should still be good even for upstream Linux to cover more
> > > > > > scenarios where we shouldn't need to do extra page faults on the small
> > > > > > pages if the previous huge page is already written, so the dirty bit
> > > > > > optimization path underneath can cover more.
> > > > > >
> > > > >
> > > > > So as said by Kirill NAK you are not looking at the right place for
> > > > > your bug please first apply the below patch and read my analysis in
> > > > > my last reply.
> > > >
> > > > Just to be clear you are trying to fix a userspace bug that is hidden
> > > > for non THP pages by a kernel space bug inside userfaultfd by making
> > > > the kernel space bug of userfaultfd buggy for THP too.
> > > >
> > > >
> > > > >
> > > > > Below patch fix userfaultfd bug. I am not posting it as it is on a
> > > > > branch and i am not sure when Andrea plan to post. Andrea feel free
> > > > > to squash that fix.
> > > > >
> > > > >
> > > > > From 35cdb30afa86424c2b9f23c0982afa6731be961c Mon Sep 17 00:00:00 2001
> > > > > From: =?UTF-8?q?J=C3=A9r=C3=B4me=20Glisse?= <jglisse@xxxxxxxxxx>
> > > > > Date: Wed, 12 Sep 2018 08:58:33 -0400
> > > > > Subject: [PATCH] userfaultfd: do not set dirty accountable when changing
> > > > > protection
> > > > > MIME-Version: 1.0
> > > > > Content-Type: text/plain; charset=UTF-8
> > > > > Content-Transfer-Encoding: 8bit
> > > > >
> > > > > mwriteprotect_range() has nothing to do with the dirty accountable
> > > > > optimization so do not set it as it opens a door for userspace to
> > > > > unwrite protect pages in a range that is write protected ie the vma
> > > > > !(vm_flags & VM_WRITE).
> > > > >
> > > > > Signed-off-by: JÃrÃme Glisse <jglisse@xxxxxxxxxx>
> > > > > ---
> > > > > mm/userfaultfd.c | 2 +-
> > > > > 1 file changed, 1 insertion(+), 1 deletion(-)
> > > > >
> > > > > diff --git a/mm/userfaultfd.c b/mm/userfaultfd.c
> > > > > index a0379c5ffa7c..59db1ce48fa0 100644
> > > > > --- a/mm/userfaultfd.c
> > > > > +++ b/mm/userfaultfd.c
> > > > > @@ -632,7 +632,7 @@ int mwriteprotect_range(struct mm_struct *dst_mm, unsigned long start,
> > > > > newprot = vm_get_page_prot(dst_vma->vm_flags);
> > > > >
> > > > > change_protection(dst_vma, start, start + len, newprot,
> > > > > - !enable_wp, 0);
> > > > > + false, 0);
> > > > >
> > > > > err = 0;
> > > > > out_unlock:
> > >
> > > Hi, Jerome,
> > >
> > > I tried your patch, unluckily it didn't work just like when not
> > > applied:
> > >
> > > Sep 13 15:16:52 px-ws kernel: FAULT_FLAG_ALLOW_RETRY missing 71
> > > Sep 13 15:16:52 px-ws kernel: CPU: 5 PID: 1625 Comm: qemu-system-x86 Not tainted 4.19.0-rc2+ #31
> > > Sep 13 15:16:52 px-ws kernel: Hardware name: LENOVO ThinkCentre M8500t-N000/SHARKBAY, BIOS FBKTC6AUS 06/22/2016
> > > Sep 13 15:16:52 px-ws kernel: Call Trace:
> > > Sep 13 15:16:52 px-ws kernel: dump_stack+0x5c/0x7b
> > > Sep 13 15:16:52 px-ws kernel: handle_userfault+0x4b5/0x780
> > > Sep 13 15:16:52 px-ws kernel: ? userfaultfd_ctx_put+0xb0/0xb0
> > > Sep 13 15:16:52 px-ws kernel: do_wp_page+0x1bd/0x5a0
> > > Sep 13 15:16:52 px-ws kernel: __handle_mm_fault+0x7f9/0x1250
> > > Sep 13 15:16:52 px-ws kernel: handle_mm_fault+0xfc/0x1f0
> > > Sep 13 15:16:52 px-ws kernel: __do_page_fault+0x255/0x520
> > > Sep 13 15:16:52 px-ws kernel: do_page_fault+0x32/0x110
> > > Sep 13 15:16:52 px-ws kernel: ? page_fault+0x8/0x30
> > > Sep 13 15:16:52 px-ws kernel: page_fault+0x1e/0x30
> > > Sep 13 15:16:52 px-ws kernel: RIP: 0033:0x7f2a9d3254e0
> > > Sep 13 15:16:52 px-ws kernel: Code: 73 01 c1 ef 07 48 81 e6 00 f0 ff ff 81 e7 e0 1f 00 00 49 8d bc 3e 40 57 00 00 48 3b 37 48 8b f3 0f 85 a4 01 00 00 48 03 77 10 <66> 89 06f
> > > Sep 13 15:16:52 px-ws kernel: RSP: 002b:00007f2ab1aae390 EFLAGS: 00010202
> > > Sep 13 15:16:52 px-ws kernel: RAX: 0000000000000246 RBX: 0000000000001ff2 RCX: 0000000000000031
> > > Sep 13 15:16:52 px-ws kernel: RDX: ffffffffffac9604 RSI: 00007f2a53e01ff2 RDI: 000055a98fa049c0
> > > Sep 13 15:16:52 px-ws kernel: RBP: 0000000000001ff4 R08: 0000000000000000 R09: 0000000000000002
> > > Sep 13 15:16:52 px-ws kernel: R10: 0000000000000000 R11: 00007f2a98201030 R12: 0000000000001ff2
> > > Sep 13 15:16:52 px-ws kernel: R13: 0000000000000000 R14: 000055a98f9ff260 R15: 00007f2ab1aaf700
> > >
> > > In case you'd like to try, here's the QEMU binary I'm testing:
> > >
> > > https://github.com/xzpeter/qemu/tree/peter-userfault-wp-test
> > >
> > > It write protects the whole system when received HMP command "info
> > > status" (I hacked that command for simplicity; it's of course not used
> > > for that...).
> > >
> > > Would you please help me understand how your patch could resolve the
> > > wp page fault from userspace if not with dirty_accountable set in the
> > > uffd-wp world (sorry for asking a question that is related to a custom
> > > tree, but finally it'll be targeted at upstream after all)? I asked
> > > this question in my previous reply to you in v1 but you didn't
> > > respond. I'd be glad to test any of your further patches if you can
> > > help solve the problem, but I'd also appreciate if you could explain
> > > it a bit on how it work since again I didn't see why it could work:
> > > again, if without that dirty_accountable set then IMO we will never
> > > setup _PAGE_WRITE for page entries and IMHO that's needed for
> > > resolving the page fault for uffd-wp tree.
> >
> > I missed that reply and forgot about PAGE_COPY ... So below is
> > what i believe a proper fix for your issue:
> >
>
> Below is a slightly better one to avoid mkwrite on COW page but it is
> still kind of ugly to do that in those function maybe adding a new helper
> would be a better way dunno. Anyway untested but it is better than trying
> to set pte dirty.
>
>
> From 8333ac9de1c54fe858f6f22dd21a2e14eebccd0e Mon Sep 17 00:00:00 2001
> From: =?UTF-8?q?J=C3=A9r=C3=B4me=20Glisse?= <jglisse@xxxxxxxxxx>
> Date: Thu, 13 Sep 2018 10:16:30 -0400
> Subject: [PATCH] mm/mprotect: add a mkwrite paramater to change_protection()
> MIME-Version: 1.0
> Content-Type: text/plain; charset=UTF-8
> Content-Transfer-Encoding: 8bit
>
> The mkwrite parameter allow to change read only pte to write one which
> is needed by userfaultfd to un-write-protect after a fault have been
> handled.
>
> Signed-off-by: JÃrÃme Glisse <jglisse@xxxxxxxxxx>
> ---
> include/linux/huge_mm.h | 2 +-
> include/linux/mm.h | 3 +-
> mm/huge_memory.c | 32 +++++++++++++++++++--
> mm/mempolicy.c | 2 +-
> mm/mprotect.c | 61 +++++++++++++++++++++++++++++------------
> mm/userfaultfd.c | 2 +-
> 6 files changed, 79 insertions(+), 23 deletions(-)
>
> diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h
> index a8a126259bc4..b51ff7f8e65c 100644
> --- a/include/linux/huge_mm.h
> +++ b/include/linux/huge_mm.h
> @@ -45,7 +45,7 @@ extern bool move_huge_pmd(struct vm_area_struct *vma, unsigned long old_addr,
> pmd_t *old_pmd, pmd_t *new_pmd, bool *need_flush);
> extern int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
> unsigned long addr, pgprot_t newprot,
> - int prot_numa);
> + int prot_numa, bool mkwrite);
> int vmf_insert_pfn_pmd(struct vm_area_struct *vma, unsigned long addr,
> pmd_t *pmd, pfn_t pfn, bool write);
> int vmf_insert_pfn_pud(struct vm_area_struct *vma, unsigned long addr,
> diff --git a/include/linux/mm.h b/include/linux/mm.h
> index 5d5c7fd07dc0..2bbf3e33bf9e 100644
> --- a/include/linux/mm.h
> +++ b/include/linux/mm.h
> @@ -1492,7 +1492,8 @@ extern unsigned long move_page_tables(struct vm_area_struct *vma,
> bool need_rmap_locks);
> extern unsigned long change_protection(struct vm_area_struct *vma, unsigned long start,
> unsigned long end, pgprot_t newprot,
> - int dirty_accountable, int prot_numa);
> + int dirty_accountable, int prot_numa,
> + bool mkwrite);
> extern int mprotect_fixup(struct vm_area_struct *vma,
> struct vm_area_struct **pprev, unsigned long start,
> unsigned long end, unsigned long newflags);
> diff --git a/mm/huge_memory.c b/mm/huge_memory.c
> index abf621aba672..3038f4bd9cce 100644
> --- a/mm/huge_memory.c
> +++ b/mm/huge_memory.c
> @@ -1842,12 +1842,13 @@ bool move_huge_pmd(struct vm_area_struct *vma, unsigned long old_addr,
> * - HPAGE_PMD_NR is protections changed and TLB flush necessary
> */
> int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
> - unsigned long addr, pgprot_t newprot, int prot_numa)
> + unsigned long addr, pgprot_t newprot, int prot_numa,
> + bool mkwrite)
> {
> struct mm_struct *mm = vma->vm_mm;
> spinlock_t *ptl;
> pmd_t entry;
> - bool preserve_write;
> + bool preserve_write, do_mkwrite = false;
> int ret;
>
> ptl = __pmd_trans_huge_lock(pmd, vma);
> @@ -1857,6 +1858,31 @@ int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
> preserve_write = prot_numa && pmd_write(*pmd);
> ret = 1;
>
> + if (mkwrite && pmd_present(*pmd) && !pmd_write(*pmd)) {
> + struct page *page = pmd_page(pmd);
> + pmd_t orig_pmd = READ_ONCE(pmd);
> +
> + VM_BUG_ON_PAGE(!PageCompound(page) || !PageHead(page), page);
> + /*
> + * We can only allow mkwrite if nobody else maps the huge page
> + * or it's part.
> + */
> + if (!trylock_page(page)) {
> + get_page(page);
> + spin_unlock(ptl);
> + lock_page(page);
> +
> + ptl = __pmd_trans_huge_lock(pmd, vma);
> + if (!ptl)
> + return 0;
> + }
> + if (pmd_same(*pmd, orig_pmd) && reuse_swap_page(page, NULL)) {
> + do_mkwrite = true;
> + }
> + unlock_page(page);
> + put_page(page);
> + }
> +
> #ifdef CONFIG_ARCH_ENABLE_THP_MIGRATION
> if (is_swap_pmd(*pmd)) {
> swp_entry_t entry = pmd_to_swp_entry(*pmd);
> @@ -1925,6 +1951,8 @@ int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
> entry = pmd_modify(entry, newprot);
> if (preserve_write)
> entry = pmd_mk_savedwrite(entry);
> + if (do_mkwrite)
> + entry = pmd_mkwrite(entry);
> ret = HPAGE_PMD_NR;
> set_pmd_at(mm, addr, pmd, entry);
> BUG_ON(vma_is_anonymous(vma) && !preserve_write && pmd_write(entry));
> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
> index 4ce44d3ff03d..2d0ee09e6b26 100644
> --- a/mm/mempolicy.c
> +++ b/mm/mempolicy.c
> @@ -579,7 +579,7 @@ unsigned long change_prot_numa(struct vm_area_struct *vma,
> {
> int nr_updated;
>
> - nr_updated = change_protection(vma, addr, end, PAGE_NONE, 0, 1);
> + nr_updated = change_protection(vma, addr, end, PAGE_NONE, 0, 1, false);
> if (nr_updated)
> count_vm_numa_events(NUMA_PTE_UPDATES, nr_updated);
>
> diff --git a/mm/mprotect.c b/mm/mprotect.c
> index 58b629bb70de..c43a29eebdaf 100644
> --- a/mm/mprotect.c
> +++ b/mm/mprotect.c
> @@ -36,7 +36,7 @@
>
> static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
> unsigned long addr, unsigned long end, pgprot_t newprot,
> - int dirty_accountable, int prot_numa)
> + int dirty_accountable, int prot_numa, bool mkwrite)
> {
> struct mm_struct *mm = vma->vm_mm;
> pte_t *pte, oldpte;
> @@ -72,13 +72,15 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
> if (pte_present(oldpte)) {
> pte_t ptent;
> bool preserve_write = prot_numa && pte_write(oldpte);
> + bool do_mkwrite = false;
>
> /*
> * Avoid trapping faults against the zero or KSM
> * pages. See similar comment in change_huge_pmd.
> */
> - if (prot_numa) {
> + if (prot_numa || mkwrite) {
> struct page *page;
> + int tmp;
>
> page = vm_normal_page(vma, addr, oldpte);
> if (!page || PageKsm(page))
> @@ -94,6 +96,26 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
> */
> if (target_node == page_to_nid(page))
> continue;
> +
> + if (mkwrite) {
> + if (!trylock_page(page)) {
> + pte_t orig_pte = READ_ONCE(pte);
> + get_page(page);
> + pte_unmap_unlock(pte, ptl);
> + lock_page(page);
> + pte = pte_offset_map_lock(vma->vm_mm, pmd,
> + addr, &ptl);
> + if (!pte_same(*pte, orig_pte)) {
> + unlock_page(page);
> + put_page(page);
> + continue;
> + }
> + }
> + if (reuse_swap_page(vmf->page, &tmp))
> + do_mkwrite = true;
> + unlock_page(page);
> + put_page(page);
> + }
> }
>
> ptent = ptep_modify_prot_start(mm, addr, pte);
> @@ -102,9 +124,9 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
> ptent = pte_mk_savedwrite(ptent);
>
> /* Avoid taking write faults for known dirty pages */
> - if (dirty_accountable && pte_dirty(ptent) &&
> - (pte_soft_dirty(ptent) ||
> - !(vma->vm_flags & VM_SOFTDIRTY))) {
> + if (do_mkwrite || (dirty_accountable &&
> + pte_dirty(ptent) && (pte_soft_dirty(ptent) ||
> + !(vma->vm_flags & VM_SOFTDIRTY)))) {
> ptent = pte_mkwrite(ptent);
> }
> ptep_modify_prot_commit(mm, addr, pte, ptent);
> @@ -150,7 +172,8 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
>
> static inline unsigned long change_pmd_range(struct vm_area_struct *vma,
> pud_t *pud, unsigned long addr, unsigned long end,
> - pgprot_t newprot, int dirty_accountable, int prot_numa)
> + pgprot_t newprot, int dirty_accountable, int prot_numa,
> + bool mkwrite)
> {
> pmd_t *pmd;
> struct mm_struct *mm = vma->vm_mm;
> @@ -179,7 +202,7 @@ static inline unsigned long change_pmd_range(struct vm_area_struct *vma,
> __split_huge_pmd(vma, pmd, addr, false, NULL);
> } else {
> int nr_ptes = change_huge_pmd(vma, pmd, addr,
> - newprot, prot_numa);
> + newprot, prot_numa, mkwrite);
>
> if (nr_ptes) {
> if (nr_ptes == HPAGE_PMD_NR) {
> @@ -194,7 +217,7 @@ static inline unsigned long change_pmd_range(struct vm_area_struct *vma,
> /* fall through, the trans huge pmd just split */
> }
> this_pages = change_pte_range(vma, pmd, addr, next, newprot,
> - dirty_accountable, prot_numa);
> + dirty_accountable, prot_numa, mkwrite);
> pages += this_pages;
> next:
> cond_resched();
> @@ -210,7 +233,8 @@ static inline unsigned long change_pmd_range(struct vm_area_struct *vma,
>
> static inline unsigned long change_pud_range(struct vm_area_struct *vma,
> p4d_t *p4d, unsigned long addr, unsigned long end,
> - pgprot_t newprot, int dirty_accountable, int prot_numa)
> + pgprot_t newprot, int dirty_accountable, int prot_numa,
> + bool mkwrite)
> {
> pud_t *pud;
> unsigned long next;
> @@ -222,7 +246,7 @@ static inline unsigned long change_pud_range(struct vm_area_struct *vma,
> if (pud_none_or_clear_bad(pud))
> continue;
> pages += change_pmd_range(vma, pud, addr, next, newprot,
> - dirty_accountable, prot_numa);
> + dirty_accountable, prot_numa, mkwrite);
> } while (pud++, addr = next, addr != end);
>
> return pages;
> @@ -230,7 +254,8 @@ static inline unsigned long change_pud_range(struct vm_area_struct *vma,
>
> static inline unsigned long change_p4d_range(struct vm_area_struct *vma,
> pgd_t *pgd, unsigned long addr, unsigned long end,
> - pgprot_t newprot, int dirty_accountable, int prot_numa)
> + pgprot_t newprot, int dirty_accountable, int prot_numa,
> + bool mkwrite)
> {
> p4d_t *p4d;
> unsigned long next;
> @@ -242,7 +267,7 @@ static inline unsigned long change_p4d_range(struct vm_area_struct *vma,
> if (p4d_none_or_clear_bad(p4d))
> continue;
> pages += change_pud_range(vma, p4d, addr, next, newprot,
> - dirty_accountable, prot_numa);
> + dirty_accountable, prot_numa, mkwrite);
> } while (p4d++, addr = next, addr != end);
>
> return pages;
> @@ -250,7 +275,7 @@ static inline unsigned long change_p4d_range(struct vm_area_struct *vma,
>
> static unsigned long change_protection_range(struct vm_area_struct *vma,
> unsigned long addr, unsigned long end, pgprot_t newprot,
> - int dirty_accountable, int prot_numa)
> + int dirty_accountable, int prot_numa, mkwrite)
> {
> struct mm_struct *mm = vma->vm_mm;
> pgd_t *pgd;
> @@ -267,7 +292,7 @@ static unsigned long change_protection_range(struct vm_area_struct *vma,
> if (pgd_none_or_clear_bad(pgd))
> continue;
> pages += change_p4d_range(vma, pgd, addr, next, newprot,
> - dirty_accountable, prot_numa);
> + dirty_accountable, prot_numa, mkwrite);
> } while (pgd++, addr = next, addr != end);
>
> /* Only flush the TLB if we actually modified any entries: */
> @@ -280,14 +305,16 @@ static unsigned long change_protection_range(struct vm_area_struct *vma,
>
> unsigned long change_protection(struct vm_area_struct *vma, unsigned long start,
> unsigned long end, pgprot_t newprot,
> - int dirty_accountable, int prot_numa)
> + int dirty_accountable, int prot_numa, bool mkwrite)
> {
> unsigned long pages;
>
> if (is_vm_hugetlb_page(vma))
> pages = hugetlb_change_protection(vma, start, end, newprot);
> else
> - pages = change_protection_range(vma, start, end, newprot, dirty_accountable, prot_numa);
> + pages = change_protection_range(vma, start, end, newprot,
> + dirty_accountable,
> + prot_numa, mkwrite);
>
> return pages;
> }
> @@ -366,7 +393,7 @@ mprotect_fixup(struct vm_area_struct *vma, struct vm_area_struct **pprev,
> vma_set_page_prot(vma);
>
> change_protection(vma, start, end, vma->vm_page_prot,
> - dirty_accountable, 0);
> + dirty_accountable, 0, false);
>
> /*
> * Private VM_LOCKED VMA becoming writable: trigger COW to avoid major
> diff --git a/mm/userfaultfd.c b/mm/userfaultfd.c
> index a0379c5ffa7c..c745c5d87523 100644
> --- a/mm/userfaultfd.c
> +++ b/mm/userfaultfd.c
> @@ -632,7 +632,7 @@ int mwriteprotect_range(struct mm_struct *dst_mm, unsigned long start,
> newprot = vm_get_page_prot(dst_vma->vm_flags);
>
> change_protection(dst_vma, start, start + len, newprot,
> - !enable_wp, 0);
> + 0, 0, !enable_wp);
>
> err = 0;
> out_unlock:

Hi, Jerome,

The first version worked for me but the 2nd didn't. Both will need to
be fixed up by myself to at least pass the compilation so I'm not sure
whether the 2nd patch didn't work because of my changes or your patch
is broken. Didn't spend more time to dig.

Anyway, thanks for these attempts and your help. Let me know if you
want me to test a 3rd version, or I'll just keep the 1st patch here in
my local tree together with the rest of the work (I'd say that's far
easier to understand than the previous oneliner) since it at least
fixes the thing up.

Regards,

--
Peter Xu