vmscan: drop page_mapping_inuse()
page_mapping_inuse() is a historic predicate function for pages that are about to be reclaimed or deactivated. According to it, a page is in use when it is mapped into page tables OR part of swap cache OR backing an mmapped file. This function is used in combination with page_referenced(), which checks for young bits in ptes and the page descriptor itself for the PG_referenced bit. Thus, checking for unmapped swap cache pages is meaningless as PG_referenced is not set for anonymous pages and unmapped pages do not have young ptes. The test makes no difference. Protecting file pages that are not by themselves mapped but are part of a mapped file is also a historic leftover for short-lived things like the exec() code in libc. However, the VM now does reference accounting and activation of pages at unmap time and thus the special treatment on reclaim is obsolete. This patch drops page_mapping_inuse() and switches the two callsites to use page_mapped() directly. Signed-off-by: Johannes Weiner <hannes@cmpxchg.org> Reviewed-by: Rik van Riel <riel@redhat.com> Cc: Minchan Kim <minchan.kim@gmail.com> Cc: OSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com> Cc: Lee Schermerhorn <lee.schermerhorn@hp.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
dfc8d636cd
commit
31c0569c3b
25
mm/vmscan.c
25
mm/vmscan.c
|
@ -262,27 +262,6 @@ unsigned long shrink_slab(unsigned long scanned, gfp_t gfp_mask,
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Called without lock on whether page is mapped, so answer is unstable */
|
|
||||||
static inline int page_mapping_inuse(struct page *page)
|
|
||||||
{
|
|
||||||
struct address_space *mapping;
|
|
||||||
|
|
||||||
/* Page is in somebody's page tables. */
|
|
||||||
if (page_mapped(page))
|
|
||||||
return 1;
|
|
||||||
|
|
||||||
/* Be more reluctant to reclaim swapcache than pagecache */
|
|
||||||
if (PageSwapCache(page))
|
|
||||||
return 1;
|
|
||||||
|
|
||||||
mapping = page_mapping(page);
|
|
||||||
if (!mapping)
|
|
||||||
return 0;
|
|
||||||
|
|
||||||
/* File is mmap'd by somebody? */
|
|
||||||
return mapping_mapped(mapping);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int is_page_cache_freeable(struct page *page)
|
static inline int is_page_cache_freeable(struct page *page)
|
||||||
{
|
{
|
||||||
/*
|
/*
|
||||||
|
@ -606,7 +585,7 @@ static enum page_references page_check_references(struct page *page,
|
||||||
if (vm_flags & VM_LOCKED)
|
if (vm_flags & VM_LOCKED)
|
||||||
return PAGEREF_RECLAIM;
|
return PAGEREF_RECLAIM;
|
||||||
|
|
||||||
if (page_mapping_inuse(page))
|
if (page_mapped(page))
|
||||||
return PAGEREF_ACTIVATE;
|
return PAGEREF_ACTIVATE;
|
||||||
|
|
||||||
/* Reclaim if clean, defer dirty pages to writeback */
|
/* Reclaim if clean, defer dirty pages to writeback */
|
||||||
|
@ -1381,7 +1360,7 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* page_referenced clears PageReferenced */
|
/* page_referenced clears PageReferenced */
|
||||||
if (page_mapping_inuse(page) &&
|
if (page_mapped(page) &&
|
||||||
page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) {
|
page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) {
|
||||||
nr_rotated++;
|
nr_rotated++;
|
||||||
/*
|
/*
|
||||||
|
|
Loading…
Reference in New Issue