mm/swap: convert deactivate_page() to folio_deactivate()
Deactivate_page() has already been converted to use folios, this change converts it to take in a folio argument instead of calling page_folio(). It also renames the function folio_deactivate() to be more consistent with other folio functions. [akpm@linux-foundation.org: fix left-over comments, per Yu Zhao] Link: https://lkml.kernel.org/r/20221221180848.20774-5-vishal.moola@gmail.com Signed-off-by: Vishal Moola (Oracle) <vishal.moola@gmail.com> Reviewed-by: Matthew Wilcox (Oracle) <willy@infradead.org> Reviewed-by: SeongJae Park <sj@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
f70da5ee8f
commit
5a9e34747c
|
@ -401,7 +401,7 @@ extern void lru_add_drain(void);
|
||||||
extern void lru_add_drain_cpu(int cpu);
|
extern void lru_add_drain_cpu(int cpu);
|
||||||
extern void lru_add_drain_cpu_zone(struct zone *zone);
|
extern void lru_add_drain_cpu_zone(struct zone *zone);
|
||||||
extern void lru_add_drain_all(void);
|
extern void lru_add_drain_all(void);
|
||||||
extern void deactivate_page(struct page *page);
|
void folio_deactivate(struct folio *folio);
|
||||||
void folio_mark_lazyfree(struct folio *folio);
|
void folio_mark_lazyfree(struct folio *folio);
|
||||||
extern void swap_setup(void);
|
extern void swap_setup(void);
|
||||||
|
|
||||||
|
|
|
@ -297,7 +297,7 @@ static inline unsigned long damon_pa_mark_accessed_or_deactivate(
|
||||||
if (mark_accessed)
|
if (mark_accessed)
|
||||||
folio_mark_accessed(folio);
|
folio_mark_accessed(folio);
|
||||||
else
|
else
|
||||||
deactivate_page(&folio->page);
|
folio_deactivate(folio);
|
||||||
folio_put(folio);
|
folio_put(folio);
|
||||||
applied += folio_nr_pages(folio);
|
applied += folio_nr_pages(folio);
|
||||||
}
|
}
|
||||||
|
|
|
@ -416,7 +416,7 @@ static int madvise_cold_or_pageout_pte_range(pmd_t *pmd,
|
||||||
list_add(&folio->lru, &folio_list);
|
list_add(&folio->lru, &folio_list);
|
||||||
}
|
}
|
||||||
} else
|
} else
|
||||||
deactivate_page(&folio->page);
|
folio_deactivate(folio);
|
||||||
huge_unlock:
|
huge_unlock:
|
||||||
spin_unlock(ptl);
|
spin_unlock(ptl);
|
||||||
if (pageout)
|
if (pageout)
|
||||||
|
@ -510,7 +510,7 @@ regular_folio:
|
||||||
list_add(&folio->lru, &folio_list);
|
list_add(&folio->lru, &folio_list);
|
||||||
}
|
}
|
||||||
} else
|
} else
|
||||||
deactivate_page(&folio->page);
|
folio_deactivate(folio);
|
||||||
}
|
}
|
||||||
|
|
||||||
arch_leave_lazy_mmu_mode();
|
arch_leave_lazy_mmu_mode();
|
||||||
|
|
|
@ -2846,11 +2846,11 @@ bool folio_mark_dirty(struct folio *folio)
|
||||||
|
|
||||||
if (likely(mapping)) {
|
if (likely(mapping)) {
|
||||||
/*
|
/*
|
||||||
* readahead/lru_deactivate_page could remain
|
* readahead/folio_deactivate could remain
|
||||||
* PG_readahead/PG_reclaim due to race with folio_end_writeback
|
* PG_readahead/PG_reclaim due to race with folio_end_writeback
|
||||||
* About readahead, if the folio is written, the flags would be
|
* About readahead, if the folio is written, the flags would be
|
||||||
* reset. So no problem.
|
* reset. So no problem.
|
||||||
* About lru_deactivate_page, if the folio is redirtied,
|
* About folio_deactivate, if the folio is redirtied,
|
||||||
* the flag will be reset. So no problem. but if the
|
* the flag will be reset. So no problem. but if the
|
||||||
* folio is used by readahead it will confuse readahead
|
* folio is used by readahead it will confuse readahead
|
||||||
* and make it restart the size rampup process. But it's
|
* and make it restart the size rampup process. But it's
|
||||||
|
|
14
mm/swap.c
14
mm/swap.c
|
@ -733,17 +733,15 @@ void deactivate_file_folio(struct folio *folio)
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* deactivate_page - deactivate a page
|
* folio_deactivate - deactivate a folio
|
||||||
* @page: page to deactivate
|
* @folio: folio to deactivate
|
||||||
*
|
*
|
||||||
* deactivate_page() moves @page to the inactive list if @page was on the active
|
* folio_deactivate() moves @folio to the inactive list if @folio was on the
|
||||||
* list and was not an unevictable page. This is done to accelerate the reclaim
|
* active list and was not unevictable. This is done to accelerate the
|
||||||
* of @page.
|
* reclaim of @folio.
|
||||||
*/
|
*/
|
||||||
void deactivate_page(struct page *page)
|
void folio_deactivate(struct folio *folio)
|
||||||
{
|
{
|
||||||
struct folio *folio = page_folio(page);
|
|
||||||
|
|
||||||
if (folio_test_lru(folio) && !folio_test_unevictable(folio) &&
|
if (folio_test_lru(folio) && !folio_test_unevictable(folio) &&
|
||||||
(folio_test_active(folio) || lru_gen_enabled())) {
|
(folio_test_active(folio) || lru_gen_enabled())) {
|
||||||
struct folio_batch *fbatch;
|
struct folio_batch *fbatch;
|
||||||
|
|
|
@ -1920,7 +1920,7 @@ retry:
|
||||||
!test_bit(PGDAT_DIRTY, &pgdat->flags))) {
|
!test_bit(PGDAT_DIRTY, &pgdat->flags))) {
|
||||||
/*
|
/*
|
||||||
* Immediately reclaim when written back.
|
* Immediately reclaim when written back.
|
||||||
* Similar in principle to deactivate_page()
|
* Similar in principle to folio_deactivate()
|
||||||
* except we already have the folio isolated
|
* except we already have the folio isolated
|
||||||
* and know it's dirty
|
* and know it's dirty
|
||||||
*/
|
*/
|
||||||
|
|
Loading…
Reference in New Issue