As mentioned initially it doesn't have any benchmark score impact at all (neither positive nor negative). I expect it might be beneficial for scores in e.g. reread scenarios.On 12/07/2009 09:36 AM, Christian Ehrhardt wrote:
Hi,Well, "Buffers" is the same kind of memory as "Cached", with
commit 56e49d - "vmscan: evict use-once pages first" changed behavior of
memory management quite a bit which should be fine.
But while tracking down a performance regression I was on the wrong path
for a while suspecting this patch is causing the regression.
Fortunately this was not the case, but I got some interesting data which
I couldn't explain completely and I thought maybe its worth to get it
clarified publicly in case someone else looks at similar data again :-)
All is about the increased amount of "Buffers" accounted as active while
loosing the same portion from "Cache" accounted as inactive in
/proc/meminfo.
I understand that with the patch applied there will be some more
pressure to file pages until the balance of active/inactive file pages
is reached.
But I didn't get how this prefers buffers compared to cache pages (I
assume dropping inactive before active was the case all the time so that
can't be the only difference between buffers/cache).
the only difference being that "Cached" is associated with
files, while "Buffers" is associated with a block device.
This means that "Buffers" is more likely to contain filesystem
metadata, while "Cached" is more likely to contain file data.
Not putting pressure on the active file list if there are a
large number of inactive file pages means that pages which were
accessed more than once get protected more from pages that were
only accessed once.
My guess is that "Buffers" is larger because the VM now caches
more (frequently used) filesystem metadata, at the expense of
caching less (used once) file data.
The scenario I'm running is a low memory system (256M total), that doesThis indeed sounds like the kind of workload that would only
sequential I/O with parallel iozone processes.
access the file data very infrequently, while accessing the
filesystem metadata all the time.
But I can't really see in the code where buffers are favored inYou are right that the code does not favor Buffers or Cache
comparison to cached pages - (it very probably makes sense to do so, as
they might contain e.g. the inode data about the files in cache).
over the other, but treats both kinds of pages the same.
I believe that you are just seeing the effect of code that
better protects the frequently accessed metadata from the
infrequently accessed data.
I try to explain the same thing as another word. if active list have
lots unimportant pages, the patch makes to gurard unimportant pages.
it might makes stream I/O benchmark score a bit because such workload
doesn't have the pages theat should be protected. iow, it only reduce
memory for cache.
The patch's intention is to improve real workload (i.e. stream/random I/O mixed workload).
not improve benchmark score. So, I'm interest how much decrease your
benchmark score.