Re: [PATCH V4] mm: fix use-after free of page_ext after race with memory-offline

From: Michal Hocko
Date: Thu Aug 18 2022 - 11:25:51 EST


On Thu 18-08-22 19:20:00, Charan Teja Kalla wrote:
> The below is one path where race between page_ext and offline of the
> respective memory blocks will cause use-after-free on the access of
> page_ext structure.
>
> process1 process2
> --------- ---------
> a)doing /proc/page_owner doing memory offline
> through offline_pages.
>
> b)PageBuddy check is failed
> thus proceed to get the
> page_owner information
> through page_ext access.
> page_ext = lookup_page_ext(page);
>
> migrate_pages();
> .................
> Since all pages are successfully
> migrated as part of the offline
> operation,send MEM_OFFLINE notification
> where for page_ext it calls:
> offline_page_ext()-->
> __free_page_ext()-->
> free_page_ext()-->
> vfree(ms->page_ext)
> mem_section->page_ext = NULL
>
> c) Check for the PAGE_EXT flags
> in the page_ext->flags access
> results into the use-after-free(leading
> to the translation faults).
>
> As mentioned above, there is really no synchronization between page_ext
> access and its freeing in the memory_offline.
>
> The memory offline steps(roughly) on a memory block is as below:
> 1) Isolate all the pages
> 2) while(1)
> try free the pages to buddy.(->free_list[MIGRATE_ISOLATE])
> 3) delete the pages from this buddy list.
> 4) Then free page_ext.(Note: The struct page is still alive as it is
> freed only during hot remove of the memory which frees the memmap, which
> steps the user might not perform).
>
> This design leads to the state where struct page is alive but the struct
> page_ext is freed, where the later is ideally part of the former which
> just representing the page_flags (check [3] for why this design is
> chosen).
>
> The above mentioned race is just one example __but the problem persists
> in the other paths too involving page_ext->flags access(eg:
> page_is_idle())__.
>
> Fix all the paths where offline races with page_ext access by
> maintaining synchronization with rcu lock and is achieved in 3 steps:
> 1) Invalidate all the page_ext's of the sections of a memory block by
> storing a flag in the LSB of mem_section->page_ext.
>
> 2) Wait till all the existing readers to finish working with the
> ->page_ext's with synchronize_rcu(). Any parallel process that starts
> after this call will not get page_ext, through lookup_page_ext(), for
> the block parallel offline operation is being performed.
>
> 3) Now safely free all sections ->page_ext's of the block on which
> offline operation is being performed.
>
> Note: If synchronize_rcu() takes time then optimizations can be done in
> this path through call_rcu()[2].
>
> Thanks to David Hildenbrand for his views/suggestions on the initial
> discussion[1] and Pavan kondeti for various inputs on this patch.
>
> [1] https://lore.kernel.org/linux-mm/59edde13-4167-8550-86f0-11fc67882107@xxxxxxxxxxx/
> [2] https://lore.kernel.org/all/a26ce299-aed1-b8ad-711e-a49e82bdd180@xxxxxxxxxxx/T/#u
> [3] https://lore.kernel.org/all/6fa6b7aa-731e-891c-3efb-a03d6a700efa@xxxxxxxxxx/
>
> Suggested-by: David Hildenbrand <david@xxxxxxxxxx>
> Suggested-by: Michal Hocko <mhocko@xxxxxxxx>
> Signed-off-by: Charan Teja Kalla <quic_charante@xxxxxxxxxxx>

I hope I haven't really missed anything but it seems all my review
feedback/concerns have been addressed. Thanks!

Acked-by: Michal Hocko <mhocko@xxxxxxxx>

> ---
> Changes in V4:
> o Updated the documentation for page_ext_get/put() -- Matthew/Michal
> o Changed the page_ext_put() interface -- Michal
> o page_ext_get/put() is called only once in set_page_owner -- vlastimil/Michal
> o Updated the commit text
>
> Changes in V3:
> o Exposed page_ext_get/put() and hid lookup_page_ext to get page_ext information.
> o Converted the call sites to use single interface i.e.page_ext_get/put().
> o Placed rcu_lock held checks where required.
> o Improved the commit message.
> o https://lore.kernel.org/all/1660056403-20894-1-git-send-email-quic_charante@xxxxxxxxxxx/
>
> Changes in V2:
> o Use only page_ext_get/put() to get the page_ext in the
> required paths. Add proper comments for them.
> o Use synchronize_rcu() only once instead of calling it for
> every mem_section::page_ext of a memory block.
> o Free'd page_ext in 3 steps of invalidate, wait till all the
> users are finished using and then finally free page_ext.
> o https://lore.kernel.org/all/1658931303-17024-1-git-send-email-quic_charante@xxxxxxxxxxx/
>
> Changes in V1:
> o Used the RCU lock while accessing the page_ext in the paths that
> can race with the memory offline operation.
> o Introduced (get|put)_page_ext() function to get the page_ext of page.
> o https://lore.kernel.org/all/1657810063-28938-1-git-send-email-quic_charante@xxxxxxxxxxx/
> include/linux/page_ext.h | 17 +++++---
> include/linux/page_idle.h | 34 ++++++++++-----
> mm/page_ext.c | 103 +++++++++++++++++++++++++++++++++++++++++++---
> mm/page_owner.c | 73 +++++++++++++++++++++++---------
> mm/page_table_check.c | 10 +++--
> 5 files changed, 192 insertions(+), 45 deletions(-)
>
> diff --git a/include/linux/page_ext.h b/include/linux/page_ext.h
> index fabb2e1..ed27198 100644
> --- a/include/linux/page_ext.h
> +++ b/include/linux/page_ext.h
> @@ -55,7 +55,8 @@ static inline void page_ext_init(void)
> }
> #endif
>
> -struct page_ext *lookup_page_ext(const struct page *page);
> +extern struct page_ext *page_ext_get(struct page *page);
> +extern void page_ext_put(struct page_ext *page_ext);
>
> static inline struct page_ext *page_ext_next(struct page_ext *curr)
> {
> @@ -71,11 +72,6 @@ static inline void pgdat_page_ext_init(struct pglist_data *pgdat)
> {
> }
>
> -static inline struct page_ext *lookup_page_ext(const struct page *page)
> -{
> - return NULL;
> -}
> -
> static inline void page_ext_init(void)
> {
> }
> @@ -87,5 +83,14 @@ static inline void page_ext_init_flatmem_late(void)
> static inline void page_ext_init_flatmem(void)
> {
> }
> +
> +static inline struct page_ext *page_ext_get(struct page *page)
> +{
> + return NULL;
> +}
> +
> +static inline void page_ext_put(struct page_ext *page_ext)
> +{
> +}
> #endif /* CONFIG_PAGE_EXTENSION */
> #endif /* __LINUX_PAGE_EXT_H */
> diff --git a/include/linux/page_idle.h b/include/linux/page_idle.h
> index 4663dfe..5cb7bd2 100644
> --- a/include/linux/page_idle.h
> +++ b/include/linux/page_idle.h
> @@ -13,65 +13,79 @@
> * If there is not enough space to store Idle and Young bits in page flags, use
> * page ext flags instead.
> */
> -
> static inline bool folio_test_young(struct folio *folio)
> {
> - struct page_ext *page_ext = lookup_page_ext(&folio->page);
> + struct page_ext *page_ext = page_ext_get(&folio->page);
> + bool page_young;
>
> if (unlikely(!page_ext))
> return false;
>
> - return test_bit(PAGE_EXT_YOUNG, &page_ext->flags);
> + page_young = test_bit(PAGE_EXT_YOUNG, &page_ext->flags);
> + page_ext_put(page_ext);
> +
> + return page_young;
> }
>
> static inline void folio_set_young(struct folio *folio)
> {
> - struct page_ext *page_ext = lookup_page_ext(&folio->page);
> + struct page_ext *page_ext = page_ext_get(&folio->page);
>
> if (unlikely(!page_ext))
> return;
>
> set_bit(PAGE_EXT_YOUNG, &page_ext->flags);
> + page_ext_put(page_ext);
> }
>
> static inline bool folio_test_clear_young(struct folio *folio)
> {
> - struct page_ext *page_ext = lookup_page_ext(&folio->page);
> + struct page_ext *page_ext = page_ext_get(&folio->page);
> + bool page_young;
>
> if (unlikely(!page_ext))
> return false;
>
> - return test_and_clear_bit(PAGE_EXT_YOUNG, &page_ext->flags);
> + page_young = test_and_clear_bit(PAGE_EXT_YOUNG, &page_ext->flags);
> + page_ext_put(page_ext);
> +
> + return page_young;
> }
>
> static inline bool folio_test_idle(struct folio *folio)
> {
> - struct page_ext *page_ext = lookup_page_ext(&folio->page);
> + struct page_ext *page_ext = page_ext_get(&folio->page);
> + bool page_idle;
>
> if (unlikely(!page_ext))
> return false;
>
> - return test_bit(PAGE_EXT_IDLE, &page_ext->flags);
> + page_idle = test_bit(PAGE_EXT_IDLE, &page_ext->flags);
> + page_ext_put(page_ext);
> +
> + return page_idle;
> }
>
> static inline void folio_set_idle(struct folio *folio)
> {
> - struct page_ext *page_ext = lookup_page_ext(&folio->page);
> + struct page_ext *page_ext = page_ext_get(&folio->page);
>
> if (unlikely(!page_ext))
> return;
>
> set_bit(PAGE_EXT_IDLE, &page_ext->flags);
> + page_ext_put(page_ext);
> }
>
> static inline void folio_clear_idle(struct folio *folio)
> {
> - struct page_ext *page_ext = lookup_page_ext(&folio->page);
> + struct page_ext *page_ext = page_ext_get(&folio->page);
>
> if (unlikely(!page_ext))
> return;
>
> clear_bit(PAGE_EXT_IDLE, &page_ext->flags);
> + page_ext_put(page_ext);
> }
> #endif /* !CONFIG_64BIT */
>
> diff --git a/mm/page_ext.c b/mm/page_ext.c
> index 3dc715d..6246f19 100644
> --- a/mm/page_ext.c
> +++ b/mm/page_ext.c
> @@ -9,6 +9,7 @@
> #include <linux/page_owner.h>
> #include <linux/page_idle.h>
> #include <linux/page_table_check.h>
> +#include <linux/rcupdate.h>
>
> /*
> * struct page extension
> @@ -59,6 +60,10 @@
> * can utilize this callback to initialize the state of it correctly.
> */
>
> +#ifdef CONFIG_SPARSEMEM
> +#define PAGE_EXT_INVALID (0x1)
> +#endif
> +
> #if defined(CONFIG_PAGE_IDLE_FLAG) && !defined(CONFIG_64BIT)
> static bool need_page_idle(void)
> {
> @@ -84,6 +89,7 @@ static struct page_ext_operations *page_ext_ops[] __initdata = {
> unsigned long page_ext_size = sizeof(struct page_ext);
>
> static unsigned long total_usage;
> +static struct page_ext *lookup_page_ext(const struct page *page);
>
> static bool __init invoke_need_callbacks(void)
> {
> @@ -125,6 +131,48 @@ static inline struct page_ext *get_entry(void *base, unsigned long index)
> return base + page_ext_size * index;
> }
>
> +/**
> + * page_ext_get() - Get the extended information for a page.
> + * @page: The page we're interested in.
> + *
> + * Ensures that the page_ext will remain valid until page_ext_put()
> + * is called.
> + *
> + * Return: NULL if no page_ext exists for this page.
> + * Context: Any context. Caller may not sleep until they have called
> + * page_ext_put().
> + */
> +struct page_ext *page_ext_get(struct page *page)
> +{
> + struct page_ext *page_ext;
> +
> + rcu_read_lock();
> + page_ext = lookup_page_ext(page);
> + if (!page_ext) {
> + rcu_read_unlock();
> + return NULL;
> + }
> +
> + return page_ext;
> +}
> +
> +/**
> + * page_ext_put() - Working with page extended information is done.
> + * @page_ext - Page extended information received from page_ext_get().
> + *
> + * The page extended information of the page may not be valid after this
> + * function is called.
> + *
> + * Return: None.
> + * Context: Any context with corresponding page_ext_get() is called.
> + */
> +void page_ext_put(struct page_ext *page_ext)
> +{
> + if (unlikely(!page_ext))
> + return;
> +
> + rcu_read_unlock();
> +}
> #ifndef CONFIG_SPARSEMEM
>
>
> @@ -133,12 +181,13 @@ void __meminit pgdat_page_ext_init(struct pglist_data *pgdat)
> pgdat->node_page_ext = NULL;
> }
>
> -struct page_ext *lookup_page_ext(const struct page *page)
> +static struct page_ext *lookup_page_ext(const struct page *page)
> {
> unsigned long pfn = page_to_pfn(page);
> unsigned long index;
> struct page_ext *base;
>
> + WARN_ON_ONCE(!rcu_read_lock_held());
> base = NODE_DATA(page_to_nid(page))->node_page_ext;
> /*
> * The sanity checks the page allocator does upon freeing a
> @@ -206,20 +255,27 @@ void __init page_ext_init_flatmem(void)
> }
>
> #else /* CONFIG_SPARSEMEM */
> +static bool page_ext_invalid(struct page_ext *page_ext)
> +{
> + return !page_ext || (((unsigned long)page_ext & PAGE_EXT_INVALID) == PAGE_EXT_INVALID);
> +}
>
> -struct page_ext *lookup_page_ext(const struct page *page)
> +static struct page_ext *lookup_page_ext(const struct page *page)
> {
> unsigned long pfn = page_to_pfn(page);
> struct mem_section *section = __pfn_to_section(pfn);
> + struct page_ext *page_ext = READ_ONCE(section->page_ext);
> +
> + WARN_ON_ONCE(!rcu_read_lock_held());
> /*
> * The sanity checks the page allocator does upon freeing a
> * page can reach here before the page_ext arrays are
> * allocated when feeding a range of pages to the allocator
> * for the first time during bootup or memory hotplug.
> */
> - if (!section->page_ext)
> + if (page_ext_invalid(page_ext))
> return NULL;
> - return get_entry(section->page_ext, pfn);
> + return get_entry(page_ext, pfn);
> }
>
> static void *__meminit alloc_page_ext(size_t size, int nid)
> @@ -298,9 +354,30 @@ static void __free_page_ext(unsigned long pfn)
> ms = __pfn_to_section(pfn);
> if (!ms || !ms->page_ext)
> return;
> - base = get_entry(ms->page_ext, pfn);
> +
> + base = READ_ONCE(ms->page_ext);
> + /*
> + * page_ext here can be valid while doing the roll back
> + * operation in online_page_ext().
> + */
> + if (page_ext_invalid(base))
> + base = (void *)base - PAGE_EXT_INVALID;
> + WRITE_ONCE(ms->page_ext, NULL);
> +
> + base = get_entry(base, pfn);
> free_page_ext(base);
> - ms->page_ext = NULL;
> +}
> +
> +static void __invalidate_page_ext(unsigned long pfn)
> +{
> + struct mem_section *ms;
> + void *val;
> +
> + ms = __pfn_to_section(pfn);
> + if (!ms || !ms->page_ext)
> + return;
> + val = (void *)ms->page_ext + PAGE_EXT_INVALID;
> + WRITE_ONCE(ms->page_ext, val);
> }
>
> static int __meminit online_page_ext(unsigned long start_pfn,
> @@ -343,6 +420,20 @@ static int __meminit offline_page_ext(unsigned long start_pfn,
> start = SECTION_ALIGN_DOWN(start_pfn);
> end = SECTION_ALIGN_UP(start_pfn + nr_pages);
>
> + /*
> + * Freeing of page_ext is done in 3 steps to avoid
> + * use-after-free of it:
> + * 1) Traverse all the sections and mark their page_ext
> + * as invalid.
> + * 2) Wait for all the existing users of page_ext who
> + * started before invalidation to finish.
> + * 3) Free the page_ext.
> + */
> + for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
> + __invalidate_page_ext(pfn);
> +
> + synchronize_rcu();
> +
> for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
> __free_page_ext(pfn);
> return 0;
> diff --git a/mm/page_owner.c b/mm/page_owner.c
> index e4c6f3f..6dbad28 100644
> --- a/mm/page_owner.c
> +++ b/mm/page_owner.c
> @@ -141,7 +141,7 @@ void __reset_page_owner(struct page *page, unsigned short order)
> struct page_owner *page_owner;
> u64 free_ts_nsec = local_clock();
>
> - page_ext = lookup_page_ext(page);
> + page_ext = page_ext_get(page);
> if (unlikely(!page_ext))
> return;
>
> @@ -153,6 +153,7 @@ void __reset_page_owner(struct page *page, unsigned short order)
> page_owner->free_ts_nsec = free_ts_nsec;
> page_ext = page_ext_next(page_ext);
> }
> + page_ext_put(page_ext);
> }
>
> static inline void __set_page_owner_handle(struct page_ext *page_ext,
> @@ -183,19 +184,21 @@ static inline void __set_page_owner_handle(struct page_ext *page_ext,
> noinline void __set_page_owner(struct page *page, unsigned short order,
> gfp_t gfp_mask)
> {
> - struct page_ext *page_ext = lookup_page_ext(page);
> + struct page_ext *page_ext;
> depot_stack_handle_t handle;
>
> + handle = save_stack(gfp_mask);
> +
> + page_ext = page_ext_get(page);
> if (unlikely(!page_ext))
> return;
> -
> - handle = save_stack(gfp_mask);
> __set_page_owner_handle(page_ext, handle, order, gfp_mask);
> + page_ext_put(page_ext);
> }
>
> void __set_page_owner_migrate_reason(struct page *page, int reason)
> {
> - struct page_ext *page_ext = lookup_page_ext(page);
> + struct page_ext *page_ext = page_ext_get(page);
> struct page_owner *page_owner;
>
> if (unlikely(!page_ext))
> @@ -203,12 +206,13 @@ void __set_page_owner_migrate_reason(struct page *page, int reason)
>
> page_owner = get_page_owner(page_ext);
> page_owner->last_migrate_reason = reason;
> + page_ext_put(page_ext);
> }
>
> void __split_page_owner(struct page *page, unsigned int nr)
> {
> int i;
> - struct page_ext *page_ext = lookup_page_ext(page);
> + struct page_ext *page_ext = page_ext_get(page);
> struct page_owner *page_owner;
>
> if (unlikely(!page_ext))
> @@ -219,17 +223,25 @@ void __split_page_owner(struct page *page, unsigned int nr)
> page_owner->order = 0;
> page_ext = page_ext_next(page_ext);
> }
> + page_ext_put(page_ext);
> }
>
> void __folio_copy_owner(struct folio *newfolio, struct folio *old)
> {
> - struct page_ext *old_ext = lookup_page_ext(&old->page);
> - struct page_ext *new_ext = lookup_page_ext(&newfolio->page);
> + struct page_ext *old_ext;
> + struct page_ext *new_ext;
> struct page_owner *old_page_owner, *new_page_owner;
>
> - if (unlikely(!old_ext || !new_ext))
> + old_ext = page_ext_get(&old->page);
> + if (unlikely(!old_ext))
> return;
>
> + new_ext = page_ext_get(&newfolio->page);
> + if (unlikely(!new_ext)) {
> + page_ext_put(old_ext);
> + return;
> + }
> +
> old_page_owner = get_page_owner(old_ext);
> new_page_owner = get_page_owner(new_ext);
> new_page_owner->order = old_page_owner->order;
> @@ -254,6 +266,8 @@ void __folio_copy_owner(struct folio *newfolio, struct folio *old)
> */
> __set_bit(PAGE_EXT_OWNER, &new_ext->flags);
> __set_bit(PAGE_EXT_OWNER_ALLOCATED, &new_ext->flags);
> + page_ext_put(new_ext);
> + page_ext_put(old_ext);
> }
>
> void pagetypeinfo_showmixedcount_print(struct seq_file *m,
> @@ -307,12 +321,12 @@ void pagetypeinfo_showmixedcount_print(struct seq_file *m,
> if (PageReserved(page))
> continue;
>
> - page_ext = lookup_page_ext(page);
> + page_ext = page_ext_get(page);
> if (unlikely(!page_ext))
> continue;
>
> if (!test_bit(PAGE_EXT_OWNER_ALLOCATED, &page_ext->flags))
> - continue;
> + goto loop;
>
> page_owner = get_page_owner(page_ext);
> page_mt = gfp_migratetype(page_owner->gfp_mask);
> @@ -323,9 +337,12 @@ void pagetypeinfo_showmixedcount_print(struct seq_file *m,
> count[pageblock_mt]++;
>
> pfn = block_end_pfn;
> + page_ext_put(page_ext);
> break;
> }
> pfn += (1UL << page_owner->order) - 1;
> +loop:
> + page_ext_put(page_ext);
> }
> }
>
> @@ -435,7 +452,7 @@ print_page_owner(char __user *buf, size_t count, unsigned long pfn,
>
> void __dump_page_owner(const struct page *page)
> {
> - struct page_ext *page_ext = lookup_page_ext(page);
> + struct page_ext *page_ext = page_ext_get((void *)page);
> struct page_owner *page_owner;
> depot_stack_handle_t handle;
> gfp_t gfp_mask;
> @@ -452,6 +469,7 @@ void __dump_page_owner(const struct page *page)
>
> if (!test_bit(PAGE_EXT_OWNER, &page_ext->flags)) {
> pr_alert("page_owner info is not present (never set?)\n");
> + page_ext_put(page_ext);
> return;
> }
>
> @@ -482,6 +500,7 @@ void __dump_page_owner(const struct page *page)
> if (page_owner->last_migrate_reason != -1)
> pr_alert("page has been migrated, last migrate reason: %s\n",
> migrate_reason_names[page_owner->last_migrate_reason]);
> + page_ext_put(page_ext);
> }
>
> static ssize_t
> @@ -508,6 +527,14 @@ read_page_owner(struct file *file, char __user *buf, size_t count, loff_t *ppos)
> /* Find an allocated page */
> for (; pfn < max_pfn; pfn++) {
> /*
> + * This temporary page_owner is required so
> + * that we can avoid the context switches while holding
> + * the rcu lock and copying the page owner information to
> + * user through copy_to_user() or GFP_KERNEL allocations.
> + */
> + struct page_owner page_owner_tmp;
> +
> + /*
> * If the new page is in a new MAX_ORDER_NR_PAGES area,
> * validate the area as existing, skip it if not
> */
> @@ -525,7 +552,7 @@ read_page_owner(struct file *file, char __user *buf, size_t count, loff_t *ppos)
> continue;
> }
>
> - page_ext = lookup_page_ext(page);
> + page_ext = page_ext_get(page);
> if (unlikely(!page_ext))
> continue;
>
> @@ -534,14 +561,14 @@ read_page_owner(struct file *file, char __user *buf, size_t count, loff_t *ppos)
> * because we don't hold the zone lock.
> */
> if (!test_bit(PAGE_EXT_OWNER, &page_ext->flags))
> - continue;
> + goto loop;
>
> /*
> * Although we do have the info about past allocation of free
> * pages, it's not relevant for current memory usage.
> */
> if (!test_bit(PAGE_EXT_OWNER_ALLOCATED, &page_ext->flags))
> - continue;
> + goto loop;
>
> page_owner = get_page_owner(page_ext);
>
> @@ -550,7 +577,7 @@ read_page_owner(struct file *file, char __user *buf, size_t count, loff_t *ppos)
> * would inflate the stats.
> */
> if (!IS_ALIGNED(pfn, 1 << page_owner->order))
> - continue;
> + goto loop;
>
> /*
> * Access to page_ext->handle isn't synchronous so we should
> @@ -558,13 +585,17 @@ read_page_owner(struct file *file, char __user *buf, size_t count, loff_t *ppos)
> */
> handle = READ_ONCE(page_owner->handle);
> if (!handle)
> - continue;
> + goto loop;
>
> /* Record the next PFN to read in the file offset */
> *ppos = (pfn - min_low_pfn) + 1;
>
> + page_owner_tmp = *page_owner;
> + page_ext_put(page_ext);
> return print_page_owner(buf, count, pfn, page,
> - page_owner, handle);
> + &page_owner_tmp, handle);
> +loop:
> + page_ext_put(page_ext);
> }
>
> return 0;
> @@ -617,18 +648,20 @@ static void init_pages_in_zone(pg_data_t *pgdat, struct zone *zone)
> if (PageReserved(page))
> continue;
>
> - page_ext = lookup_page_ext(page);
> + page_ext = page_ext_get(page);
> if (unlikely(!page_ext))
> continue;
>
> /* Maybe overlapping zone */
> if (test_bit(PAGE_EXT_OWNER, &page_ext->flags))
> - continue;
> + goto loop;
>
> /* Found early allocated page */
> __set_page_owner_handle(page_ext, early_handle,
> 0, 0);
> count++;
> +loop:
> + page_ext_put(page_ext);
> }
> cond_resched();
> }
> diff --git a/mm/page_table_check.c b/mm/page_table_check.c
> index e206274..903db62 100644
> --- a/mm/page_table_check.c
> +++ b/mm/page_table_check.c
> @@ -68,7 +68,7 @@ static void page_table_check_clear(struct mm_struct *mm, unsigned long addr,
> return;
>
> page = pfn_to_page(pfn);
> - page_ext = lookup_page_ext(page);
> + page_ext = page_ext_get(page);
> anon = PageAnon(page);
>
> for (i = 0; i < pgcnt; i++) {
> @@ -83,6 +83,7 @@ static void page_table_check_clear(struct mm_struct *mm, unsigned long addr,
> }
> page_ext = page_ext_next(page_ext);
> }
> + page_ext_put(page_ext);
> }
>
> /*
> @@ -103,7 +104,7 @@ static void page_table_check_set(struct mm_struct *mm, unsigned long addr,
> return;
>
> page = pfn_to_page(pfn);
> - page_ext = lookup_page_ext(page);
> + page_ext = page_ext_get(page);
> anon = PageAnon(page);
>
> for (i = 0; i < pgcnt; i++) {
> @@ -118,6 +119,7 @@ static void page_table_check_set(struct mm_struct *mm, unsigned long addr,
> }
> page_ext = page_ext_next(page_ext);
> }
> + page_ext_put(page_ext);
> }
>
> /*
> @@ -126,9 +128,10 @@ static void page_table_check_set(struct mm_struct *mm, unsigned long addr,
> */
> void __page_table_check_zero(struct page *page, unsigned int order)
> {
> - struct page_ext *page_ext = lookup_page_ext(page);
> + struct page_ext *page_ext;
> unsigned long i;
>
> + page_ext = page_ext_get(page);
> BUG_ON(!page_ext);
> for (i = 0; i < (1ul << order); i++) {
> struct page_table_check *ptc = get_page_table_check(page_ext);
> @@ -137,6 +140,7 @@ void __page_table_check_zero(struct page *page, unsigned int order)
> BUG_ON(atomic_read(&ptc->file_map_count));
> page_ext = page_ext_next(page_ext);
> }
> + page_ext_put(page_ext);
> }
>
> void __page_table_check_pte_clear(struct mm_struct *mm, unsigned long addr,
> --
> 2.7.4

--
Michal Hocko
SUSE Labs