vmscan: drop page_mapping_inuse()

page_mapping_inuse() is a historic predicate function for pages that are
about to be reclaimed or deactivated.

According to it, a page is in use when it is mapped into page tables OR
part of swap cache OR backing an mmapped file.

This function is used in combination with page_referenced(), which checks
for young bits in ptes and the page descriptor itself for the
PG_referenced bit.  Thus, checking for unmapped swap cache pages is
meaningless as PG_referenced is not set for anonymous pages and unmapped
pages do not have young ptes.  The test makes no difference.

Protecting file pages that are not by themselves mapped but are part of a
mapped file is also a historic leftover for short-lived things like the
exec() code in libc.  However, the VM now does reference accounting and
activation of pages at unmap time and thus the special treatment on
reclaim is obsolete.

This patch drops page_mapping_inuse() and switches the two callsites to
use page_mapped() directly.

Signed-off-by: Johannes Weiner <hannes@cmpxchg.org>
Reviewed-by: Rik van Riel <riel@redhat.com>
Cc: Minchan Kim <minchan.kim@gmail.com>
Cc: OSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Cc: Lee Schermerhorn <lee.schermerhorn@hp.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Johannes Weiner 2010-03-05 13:42:21 -08:00 committed by Linus Torvalds
parent dfc8d636cd
commit 31c0569c3b
1 changed files with 2 additions and 23 deletions

View File

@ -262,27 +262,6 @@ unsigned long shrink_slab(unsigned long scanned, gfp_t gfp_mask,
return ret; return ret;
} }
/* Called without lock on whether page is mapped, so answer is unstable */
static inline int page_mapping_inuse(struct page *page)
{
struct address_space *mapping;
/* Page is in somebody's page tables. */
if (page_mapped(page))
return 1;
/* Be more reluctant to reclaim swapcache than pagecache */
if (PageSwapCache(page))
return 1;
mapping = page_mapping(page);
if (!mapping)
return 0;
/* File is mmap'd by somebody? */
return mapping_mapped(mapping);
}
static inline int is_page_cache_freeable(struct page *page) static inline int is_page_cache_freeable(struct page *page)
{ {
/* /*
@ -606,7 +585,7 @@ static enum page_references page_check_references(struct page *page,
if (vm_flags & VM_LOCKED) if (vm_flags & VM_LOCKED)
return PAGEREF_RECLAIM; return PAGEREF_RECLAIM;
if (page_mapping_inuse(page)) if (page_mapped(page))
return PAGEREF_ACTIVATE; return PAGEREF_ACTIVATE;
/* Reclaim if clean, defer dirty pages to writeback */ /* Reclaim if clean, defer dirty pages to writeback */
@ -1381,7 +1360,7 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
} }
/* page_referenced clears PageReferenced */ /* page_referenced clears PageReferenced */
if (page_mapping_inuse(page) && if (page_mapped(page) &&
page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) { page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) {
nr_rotated++; nr_rotated++;
/* /*