Re: [RFC 1/3] mm, oom: refactor oom detection

From: Hillf Danton
Date: Fri Oct 30 2015 - 23:57:54 EST


> On Fri 30-10-15 09:36:26, Michal Hocko wrote:
> > On Fri 30-10-15 12:10:15, Hillf Danton wrote:
> > [...]
> > > > + for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, ac->nodemask) {
> > > > + unsigned long free = zone_page_state(zone, NR_FREE_PAGES);
> > > > + unsigned long reclaimable;
> > > > + unsigned long target;
> > > > +
> > > > + reclaimable = zone_reclaimable_pages(zone) +
> > > > + zone_page_state(zone, NR_ISOLATED_FILE) +
> > > > + zone_page_state(zone, NR_ISOLATED_ANON);
> > > > + target = reclaimable;
> > > > + target -= stall_backoff * (1 + target/MAX_STALL_BACKOFF);
> > >
> > > target = reclaimable - stall_backoff * (1 + target/MAX_STALL_BACKOFF);
> > > = reclaimable - stall_backoff - stall_backoff * (target/MAX_STALL_BACKOFF);
> > >
> > > then the first stall_backoff looks unreasonable.
> >
> > First stall_backoff is off by 1 but that shouldn't make any difference.
> >
> > > I guess you mean
> > > target = reclaimable - target * (stall_backoff/MAX_STALL_BACKOFF);
> > > = reclaimable - stall_back * (target/MAX_STALL_BACKOFF);
> >
> > No the reason I used the bias is to converge for MAX_STALL_BACKOFF. If
> > you have target which is not divisible by MAX_STALL_BACKOFF then the
> > rounding would get target > 0 and so we wouldn't converge. With the +1
> > you underflow which is MAX_STALL_BACKOFF in maximum which should be
> > fixed up by the free memory. Maybe a check for free < MAX_STALL_BACKOFF
> > would be good but I didn't get that far with this.
>
> I've ended up with the following after all. It uses ceiling for the
> division this should be underflow safe albeit less readable (at least
> for me).

Looks good, thanks.

Acked-by: Hillf Danton <hillf.zj@xxxxxxxxxxxxxxx>

> ---
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 0dc1ca9b1219..c9a4e62f234e 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -3176,7 +3176,7 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> zone_page_state(zone, NR_ISOLATED_FILE) +
> zone_page_state(zone, NR_ISOLATED_ANON);
> target = reclaimable;
> - target -= stall_backoff * (1 + target/MAX_STALL_BACKOFF);
> + target -= (stall_backoff * target + MAX_STALL_BACKOFF - 1) / MAX_STALL_BACKOFF;
> target += free;
>
> /*
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index bc14217acd47..0b3ec972ec7a 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -2672,7 +2672,6 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
> int initial_priority = sc->priority;
> unsigned long total_scanned = 0;
> unsigned long writeback_threshold;
> - bool zones_reclaimable;
> retry:
> delayacct_freepages_start();
>
> @@ -2683,7 +2682,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
> vmpressure_prio(sc->gfp_mask, sc->target_mem_cgroup,
> sc->priority);
> sc->nr_scanned = 0;
> - zones_reclaimable = shrink_zones(zonelist, sc);
> + shrink_zones(zonelist, sc);
>
> total_scanned += sc->nr_scanned;
> if (sc->nr_reclaimed >= sc->nr_to_reclaim)
> --
> Michal Hocko
> SUSE Labs

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/