Re: [PATCH 3/5] userfualtfd: Replace lru_cache functions with folio_add functions

From: Matthew Wilcox
Date: Tue Nov 01 2022 - 14:31:31 EST


On Tue, Nov 01, 2022 at 10:53:24AM -0700, Vishal Moola (Oracle) wrote:
> Replaces lru_cache_add() and lru_cache_add_inactive_or_unevictable()
> with folio_add_lru() and folio_add_lru_vma(). This is in preparation for
> the removal of lru_cache_add().

Ummmmm. Reviewing this patch reveals a bug (not introduced by your
patch). Look:

mfill_atomic_install_pte:
bool page_in_cache = page->mapping;

mcontinue_atomic_pte:
ret = shmem_get_folio(inode, pgoff, &folio, SGP_NOALLOC);
...
page = folio_file_page(folio, pgoff);
ret = mfill_atomic_install_pte(dst_mm, dst_pmd, dst_vma, dst_addr,
page, false, wp_copy);

That says pretty plainly that mfill_atomic_install_pte() can be passed
a tail page from shmem, and if it is ...

if (page_in_cache) {
...
} else {
page_add_new_anon_rmap(page, dst_vma, dst_addr);
lru_cache_add_inactive_or_unevictable(page, dst_vma);
}

it'll get put on the rmap as an anon page!

> Signed-off-by: Vishal Moola (Oracle) <vishal.moola@xxxxxxxxx>
> ---
> mm/userfaultfd.c | 6 ++++--
> 1 file changed, 4 insertions(+), 2 deletions(-)
>
> diff --git a/mm/userfaultfd.c b/mm/userfaultfd.c
> index e24e8a47ce8a..2560973b00d8 100644
> --- a/mm/userfaultfd.c
> +++ b/mm/userfaultfd.c
> @@ -66,6 +66,7 @@ int mfill_atomic_install_pte(struct mm_struct *dst_mm, pmd_t *dst_pmd,
> bool vm_shared = dst_vma->vm_flags & VM_SHARED;
> bool page_in_cache = page->mapping;
> spinlock_t *ptl;
> + struct folio *folio;
> struct inode *inode;
> pgoff_t offset, max_off;
>
> @@ -113,14 +114,15 @@ int mfill_atomic_install_pte(struct mm_struct *dst_mm, pmd_t *dst_pmd,
> if (!pte_none_mostly(*dst_pte))
> goto out_unlock;
>
> + folio = page_folio(page);
> if (page_in_cache) {
> /* Usually, cache pages are already added to LRU */
> if (newly_allocated)
> - lru_cache_add(page);
> + folio_add_lru(folio);
> page_add_file_rmap(page, dst_vma, false);
> } else {
> page_add_new_anon_rmap(page, dst_vma, dst_addr);
> - lru_cache_add_inactive_or_unevictable(page, dst_vma);
> + folio_add_lru_vma(folio, dst_vma);
> }
>
> /*
> --
> 2.38.1
>
>