[PATCH v2 6/8] mm/isolation: factor out pre/post logic on set/unset_migratetype_isolate()

From: Joonsoo Kim
Date: Wed Aug 06 2014 - 03:13:26 EST


Current isolation logic isolates each pageblock individually.
This causes freepage counting problem when page with pageblock order is
merged with other page on different buddy list. To prevent it, we should
handle whole range at one time in start_isolate_page_range(). This patch
is preparation of that work.

Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@xxxxxxx>
---
mm/page_isolation.c | 45 +++++++++++++++++++++++++++++----------------
1 file changed, 29 insertions(+), 16 deletions(-)

diff --git a/mm/page_isolation.c b/mm/page_isolation.c
index 898361f..b91f9ec 100644
--- a/mm/page_isolation.c
+++ b/mm/page_isolation.c
@@ -78,18 +78,14 @@ static void activate_isolated_pages(struct zone *zone, unsigned long start_pfn,
spin_unlock_irqrestore(&zone->lock, flags);
}

-int set_migratetype_isolate(struct page *page, bool skip_hwpoisoned_pages)
+static int set_migratetype_isolate_pre(struct page *page,
+ bool skip_hwpoisoned_pages)
{
- struct zone *zone;
- unsigned long flags, pfn;
+ struct zone *zone = page_zone(page);
+ unsigned long pfn;
struct memory_isolate_notify arg;
int notifier_ret;
int ret = -EBUSY;
- unsigned long nr_pages;
- int migratetype;
-
- zone = page_zone(page);
- spin_lock_irqsave(&zone->lock, flags);

pfn = page_to_pfn(page);
arg.start_pfn = pfn;
@@ -110,7 +106,7 @@ int set_migratetype_isolate(struct page *page, bool skip_hwpoisoned_pages)
notifier_ret = memory_isolate_notify(MEM_ISOLATE_COUNT, &arg);
notifier_ret = notifier_to_errno(notifier_ret);
if (notifier_ret)
- goto out;
+ return ret;
/*
* FIXME: Now, memory hotplug doesn't call shrink_slab() by itself.
* We just check MOVABLE pages.
@@ -124,10 +120,20 @@ int set_migratetype_isolate(struct page *page, bool skip_hwpoisoned_pages)
* removable-by-driver pages reported by notifier, we'll fail.
*/

-out:
- if (ret) {
+ return ret;
+}
+
+int set_migratetype_isolate(struct page *page, bool skip_hwpoisoned_pages)
+{
+ struct zone *zone = page_zone(page);
+ unsigned long flags;
+ unsigned long nr_pages;
+ int migratetype;
+
+ spin_lock_irqsave(&zone->lock, flags);
+ if (set_migratetype_isolate_pre(page, skip_hwpoisoned_pages)) {
spin_unlock_irqrestore(&zone->lock, flags);
- return ret;
+ return -EBUSY;
}

migratetype = get_pageblock_migratetype(page);
@@ -153,11 +159,20 @@ out:
return 0;
}

+static void unset_migratetype_isolate_post(struct page *page,
+ unsigned migratetype)
+{
+ struct zone *zone = page_zone(page);
+ unsigned long start_pfn, end_pfn;
+
+ start_pfn = page_to_pfn(page) & ~(pageblock_nr_pages - 1);
+ end_pfn = start_pfn + pageblock_nr_pages;
+ activate_isolated_pages(zone, start_pfn, end_pfn, migratetype);
+}
void unset_migratetype_isolate(struct page *page, unsigned migratetype)
{
struct zone *zone;
unsigned long flags, nr_pages;
- unsigned long start_pfn, end_pfn;

zone = page_zone(page);
spin_lock_irqsave(&zone->lock, flags);
@@ -174,9 +189,7 @@ void unset_migratetype_isolate(struct page *page, unsigned migratetype)
/* Freed pages will see original migratetype after this point */
kick_all_cpus_sync();

- start_pfn = page_to_pfn(page) & ~(pageblock_nr_pages - 1);
- end_pfn = start_pfn + pageblock_nr_pages;
- activate_isolated_pages(zone, start_pfn, end_pfn, migratetype);
+ unset_migratetype_isolate_post(page, migratetype);
}

static inline struct page *
--
1.7.9.5

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/