mm/memcg: Convert mem_cgroup_charge() to take a folio
Convert all callers of mem_cgroup_charge() to call page_folio() on the page they're currently passing in. Many of them will be converted to use folios themselves soon. Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Reviewed-by: Christoph Hellwig <hch@lst.de> Reviewed-by: David Howells <dhowells@redhat.com> Acked-by: Vlastimil Babka <vbabka@suse.cz>
This commit is contained in:
parent
118f287549
commit
8f425e4ed0
|
@ -694,14 +694,28 @@ static inline bool mem_cgroup_below_min(struct mem_cgroup *memcg)
|
||||||
page_counter_read(&memcg->memory);
|
page_counter_read(&memcg->memory);
|
||||||
}
|
}
|
||||||
|
|
||||||
int __mem_cgroup_charge(struct page *page, struct mm_struct *mm,
|
int __mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp);
|
||||||
gfp_t gfp_mask);
|
|
||||||
static inline int mem_cgroup_charge(struct page *page, struct mm_struct *mm,
|
/**
|
||||||
gfp_t gfp_mask)
|
* mem_cgroup_charge - Charge a newly allocated folio to a cgroup.
|
||||||
|
* @folio: Folio to charge.
|
||||||
|
* @mm: mm context of the allocating task.
|
||||||
|
* @gfp: Reclaim mode.
|
||||||
|
*
|
||||||
|
* Try to charge @folio to the memcg that @mm belongs to, reclaiming
|
||||||
|
* pages according to @gfp if necessary. If @mm is NULL, try to
|
||||||
|
* charge to the active memcg.
|
||||||
|
*
|
||||||
|
* Do not use this for folios allocated for swapin.
|
||||||
|
*
|
||||||
|
* Return: 0 on success. Otherwise, an error code is returned.
|
||||||
|
*/
|
||||||
|
static inline int mem_cgroup_charge(struct folio *folio, struct mm_struct *mm,
|
||||||
|
gfp_t gfp)
|
||||||
{
|
{
|
||||||
if (mem_cgroup_disabled())
|
if (mem_cgroup_disabled())
|
||||||
return 0;
|
return 0;
|
||||||
return __mem_cgroup_charge(page, mm, gfp_mask);
|
return __mem_cgroup_charge(folio, mm, gfp);
|
||||||
}
|
}
|
||||||
|
|
||||||
int mem_cgroup_swapin_charge_page(struct page *page, struct mm_struct *mm,
|
int mem_cgroup_swapin_charge_page(struct page *page, struct mm_struct *mm,
|
||||||
|
@ -1199,8 +1213,8 @@ static inline bool mem_cgroup_below_min(struct mem_cgroup *memcg)
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int mem_cgroup_charge(struct page *page, struct mm_struct *mm,
|
static inline int mem_cgroup_charge(struct folio *folio,
|
||||||
gfp_t gfp_mask)
|
struct mm_struct *mm, gfp_t gfp)
|
||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -167,7 +167,8 @@ static int __replace_page(struct vm_area_struct *vma, unsigned long addr,
|
||||||
addr + PAGE_SIZE);
|
addr + PAGE_SIZE);
|
||||||
|
|
||||||
if (new_page) {
|
if (new_page) {
|
||||||
err = mem_cgroup_charge(new_page, vma->vm_mm, GFP_KERNEL);
|
err = mem_cgroup_charge(page_folio(new_page), vma->vm_mm,
|
||||||
|
GFP_KERNEL);
|
||||||
if (err)
|
if (err)
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
|
@ -889,7 +889,7 @@ noinline int __add_to_page_cache_locked(struct page *page,
|
||||||
page->index = offset;
|
page->index = offset;
|
||||||
|
|
||||||
if (!huge) {
|
if (!huge) {
|
||||||
error = mem_cgroup_charge(page, NULL, gfp);
|
error = mem_cgroup_charge(page_folio(page), NULL, gfp);
|
||||||
if (error)
|
if (error)
|
||||||
goto error;
|
goto error;
|
||||||
charged = true;
|
charged = true;
|
||||||
|
|
|
@ -603,7 +603,7 @@ static vm_fault_t __do_huge_pmd_anonymous_page(struct vm_fault *vmf,
|
||||||
|
|
||||||
VM_BUG_ON_PAGE(!PageCompound(page), page);
|
VM_BUG_ON_PAGE(!PageCompound(page), page);
|
||||||
|
|
||||||
if (mem_cgroup_charge(page, vma->vm_mm, gfp)) {
|
if (mem_cgroup_charge(page_folio(page), vma->vm_mm, gfp)) {
|
||||||
put_page(page);
|
put_page(page);
|
||||||
count_vm_event(THP_FAULT_FALLBACK);
|
count_vm_event(THP_FAULT_FALLBACK);
|
||||||
count_vm_event(THP_FAULT_FALLBACK_CHARGE);
|
count_vm_event(THP_FAULT_FALLBACK_CHARGE);
|
||||||
|
|
|
@ -1087,7 +1087,7 @@ static void collapse_huge_page(struct mm_struct *mm,
|
||||||
goto out_nolock;
|
goto out_nolock;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (unlikely(mem_cgroup_charge(new_page, mm, gfp))) {
|
if (unlikely(mem_cgroup_charge(page_folio(new_page), mm, gfp))) {
|
||||||
result = SCAN_CGROUP_CHARGE_FAIL;
|
result = SCAN_CGROUP_CHARGE_FAIL;
|
||||||
goto out_nolock;
|
goto out_nolock;
|
||||||
}
|
}
|
||||||
|
@ -1658,7 +1658,7 @@ static void collapse_file(struct mm_struct *mm,
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (unlikely(mem_cgroup_charge(new_page, mm, gfp))) {
|
if (unlikely(mem_cgroup_charge(page_folio(new_page), mm, gfp))) {
|
||||||
result = SCAN_CGROUP_CHARGE_FAIL;
|
result = SCAN_CGROUP_CHARGE_FAIL;
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
3
mm/ksm.c
3
mm/ksm.c
|
@ -2578,7 +2578,8 @@ struct page *ksm_might_need_to_copy(struct page *page,
|
||||||
return page; /* let do_swap_page report the error */
|
return page; /* let do_swap_page report the error */
|
||||||
|
|
||||||
new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, address);
|
new_page = alloc_page_vma(GFP_HIGHUSER_MOVABLE, vma, address);
|
||||||
if (new_page && mem_cgroup_charge(new_page, vma->vm_mm, GFP_KERNEL)) {
|
if (new_page &&
|
||||||
|
mem_cgroup_charge(page_folio(new_page), vma->vm_mm, GFP_KERNEL)) {
|
||||||
put_page(new_page);
|
put_page(new_page);
|
||||||
new_page = NULL;
|
new_page = NULL;
|
||||||
}
|
}
|
||||||
|
|
|
@ -6660,9 +6660,9 @@ void mem_cgroup_calculate_protection(struct mem_cgroup *root,
|
||||||
atomic_long_read(&parent->memory.children_low_usage)));
|
atomic_long_read(&parent->memory.children_low_usage)));
|
||||||
}
|
}
|
||||||
|
|
||||||
static int charge_memcg(struct page *page, struct mem_cgroup *memcg, gfp_t gfp)
|
static int charge_memcg(struct folio *folio, struct mem_cgroup *memcg,
|
||||||
|
gfp_t gfp)
|
||||||
{
|
{
|
||||||
struct folio *folio = page_folio(page);
|
|
||||||
long nr_pages = folio_nr_pages(folio);
|
long nr_pages = folio_nr_pages(folio);
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
|
@ -6675,34 +6675,19 @@ static int charge_memcg(struct page *page, struct mem_cgroup *memcg, gfp_t gfp)
|
||||||
|
|
||||||
local_irq_disable();
|
local_irq_disable();
|
||||||
mem_cgroup_charge_statistics(memcg, nr_pages);
|
mem_cgroup_charge_statistics(memcg, nr_pages);
|
||||||
memcg_check_events(memcg, page_to_nid(page));
|
memcg_check_events(memcg, folio_nid(folio));
|
||||||
local_irq_enable();
|
local_irq_enable();
|
||||||
out:
|
out:
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
int __mem_cgroup_charge(struct folio *folio, struct mm_struct *mm, gfp_t gfp)
|
||||||
* __mem_cgroup_charge - charge a newly allocated page to a cgroup
|
|
||||||
* @page: page to charge
|
|
||||||
* @mm: mm context of the victim
|
|
||||||
* @gfp_mask: reclaim mode
|
|
||||||
*
|
|
||||||
* Try to charge @page to the memcg that @mm belongs to, reclaiming
|
|
||||||
* pages according to @gfp_mask if necessary. if @mm is NULL, try to
|
|
||||||
* charge to the active memcg.
|
|
||||||
*
|
|
||||||
* Do not use this for pages allocated for swapin.
|
|
||||||
*
|
|
||||||
* Returns 0 on success. Otherwise, an error code is returned.
|
|
||||||
*/
|
|
||||||
int __mem_cgroup_charge(struct page *page, struct mm_struct *mm,
|
|
||||||
gfp_t gfp_mask)
|
|
||||||
{
|
{
|
||||||
struct mem_cgroup *memcg;
|
struct mem_cgroup *memcg;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
memcg = get_mem_cgroup_from_mm(mm);
|
memcg = get_mem_cgroup_from_mm(mm);
|
||||||
ret = charge_memcg(page, memcg, gfp_mask);
|
ret = charge_memcg(folio, memcg, gfp);
|
||||||
css_put(&memcg->css);
|
css_put(&memcg->css);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
|
@ -6723,6 +6708,7 @@ int __mem_cgroup_charge(struct page *page, struct mm_struct *mm,
|
||||||
int mem_cgroup_swapin_charge_page(struct page *page, struct mm_struct *mm,
|
int mem_cgroup_swapin_charge_page(struct page *page, struct mm_struct *mm,
|
||||||
gfp_t gfp, swp_entry_t entry)
|
gfp_t gfp, swp_entry_t entry)
|
||||||
{
|
{
|
||||||
|
struct folio *folio = page_folio(page);
|
||||||
struct mem_cgroup *memcg;
|
struct mem_cgroup *memcg;
|
||||||
unsigned short id;
|
unsigned short id;
|
||||||
int ret;
|
int ret;
|
||||||
|
@ -6737,7 +6723,7 @@ int mem_cgroup_swapin_charge_page(struct page *page, struct mm_struct *mm,
|
||||||
memcg = get_mem_cgroup_from_mm(mm);
|
memcg = get_mem_cgroup_from_mm(mm);
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
|
||||||
ret = charge_memcg(page, memcg, gfp);
|
ret = charge_memcg(folio, memcg, gfp);
|
||||||
|
|
||||||
css_put(&memcg->css);
|
css_put(&memcg->css);
|
||||||
return ret;
|
return ret;
|
||||||
|
|
|
@ -990,7 +990,7 @@ page_copy_prealloc(struct mm_struct *src_mm, struct vm_area_struct *vma,
|
||||||
if (!new_page)
|
if (!new_page)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
if (mem_cgroup_charge(new_page, src_mm, GFP_KERNEL)) {
|
if (mem_cgroup_charge(page_folio(new_page), src_mm, GFP_KERNEL)) {
|
||||||
put_page(new_page);
|
put_page(new_page);
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
@ -3019,7 +3019,7 @@ static vm_fault_t wp_page_copy(struct vm_fault *vmf)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (mem_cgroup_charge(new_page, mm, GFP_KERNEL))
|
if (mem_cgroup_charge(page_folio(new_page), mm, GFP_KERNEL))
|
||||||
goto oom_free_new;
|
goto oom_free_new;
|
||||||
cgroup_throttle_swaprate(new_page, GFP_KERNEL);
|
cgroup_throttle_swaprate(new_page, GFP_KERNEL);
|
||||||
|
|
||||||
|
@ -3769,7 +3769,7 @@ static vm_fault_t do_anonymous_page(struct vm_fault *vmf)
|
||||||
if (!page)
|
if (!page)
|
||||||
goto oom;
|
goto oom;
|
||||||
|
|
||||||
if (mem_cgroup_charge(page, vma->vm_mm, GFP_KERNEL))
|
if (mem_cgroup_charge(page_folio(page), vma->vm_mm, GFP_KERNEL))
|
||||||
goto oom_free_page;
|
goto oom_free_page;
|
||||||
cgroup_throttle_swaprate(page, GFP_KERNEL);
|
cgroup_throttle_swaprate(page, GFP_KERNEL);
|
||||||
|
|
||||||
|
@ -4193,7 +4193,8 @@ static vm_fault_t do_cow_fault(struct vm_fault *vmf)
|
||||||
if (!vmf->cow_page)
|
if (!vmf->cow_page)
|
||||||
return VM_FAULT_OOM;
|
return VM_FAULT_OOM;
|
||||||
|
|
||||||
if (mem_cgroup_charge(vmf->cow_page, vma->vm_mm, GFP_KERNEL)) {
|
if (mem_cgroup_charge(page_folio(vmf->cow_page), vma->vm_mm,
|
||||||
|
GFP_KERNEL)) {
|
||||||
put_page(vmf->cow_page);
|
put_page(vmf->cow_page);
|
||||||
return VM_FAULT_OOM;
|
return VM_FAULT_OOM;
|
||||||
}
|
}
|
||||||
|
|
|
@ -2846,7 +2846,7 @@ static void migrate_vma_insert_page(struct migrate_vma *migrate,
|
||||||
|
|
||||||
if (unlikely(anon_vma_prepare(vma)))
|
if (unlikely(anon_vma_prepare(vma)))
|
||||||
goto abort;
|
goto abort;
|
||||||
if (mem_cgroup_charge(page, vma->vm_mm, GFP_KERNEL))
|
if (mem_cgroup_charge(page_folio(page), vma->vm_mm, GFP_KERNEL))
|
||||||
goto abort;
|
goto abort;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -710,7 +710,7 @@ static int shmem_add_to_page_cache(struct page *page,
|
||||||
page->index = index;
|
page->index = index;
|
||||||
|
|
||||||
if (!PageSwapCache(page)) {
|
if (!PageSwapCache(page)) {
|
||||||
error = mem_cgroup_charge(page, charge_mm, gfp);
|
error = mem_cgroup_charge(page_folio(page), charge_mm, gfp);
|
||||||
if (error) {
|
if (error) {
|
||||||
if (PageTransHuge(page)) {
|
if (PageTransHuge(page)) {
|
||||||
count_vm_event(THP_FILE_FALLBACK);
|
count_vm_event(THP_FILE_FALLBACK);
|
||||||
|
|
|
@ -164,7 +164,7 @@ static int mcopy_atomic_pte(struct mm_struct *dst_mm,
|
||||||
__SetPageUptodate(page);
|
__SetPageUptodate(page);
|
||||||
|
|
||||||
ret = -ENOMEM;
|
ret = -ENOMEM;
|
||||||
if (mem_cgroup_charge(page, dst_mm, GFP_KERNEL))
|
if (mem_cgroup_charge(page_folio(page), dst_mm, GFP_KERNEL))
|
||||||
goto out_release;
|
goto out_release;
|
||||||
|
|
||||||
ret = mfill_atomic_install_pte(dst_mm, dst_pmd, dst_vma, dst_addr,
|
ret = mfill_atomic_install_pte(dst_mm, dst_pmd, dst_vma, dst_addr,
|
||||||
|
|
Loading…
Reference in New Issue