mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2025-01-01 10:45:49 +00:00
mm: convert deferred_split_huge_page() to deferred_split_folio()
Now that both callers use a folio, pass the folio in and save a call to compound_head(). Link: https://lkml.kernel.org/r/20230111142915.1001531-28-willy@infradead.org Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
f8baa6be03
commit
f158ed6195
@ -153,8 +153,8 @@ clear where references should go after split: it will stay on the head page.
|
||||
Note that split_huge_pmd() doesn't have any limitations on refcounting:
|
||||
pmd can be split at any point and never fails.
|
||||
|
||||
Partial unmap and deferred_split_huge_page()
|
||||
============================================
|
||||
Partial unmap and deferred_split_folio()
|
||||
========================================
|
||||
|
||||
Unmapping part of THP (with munmap() or other way) is not going to free
|
||||
memory immediately. Instead, we detect that a subpage of THP is not in use
|
||||
@ -166,6 +166,6 @@ the place where we can detect partial unmap. It also might be
|
||||
counterproductive since in many cases partial unmap happens during exit(2) if
|
||||
a THP crosses a VMA boundary.
|
||||
|
||||
The function deferred_split_huge_page() is used to queue a page for splitting.
|
||||
The function deferred_split_folio() is used to queue a folio for splitting.
|
||||
The splitting itself will happen when we get memory pressure via shrinker
|
||||
interface.
|
||||
|
@ -187,7 +187,7 @@ static inline int split_huge_page(struct page *page)
|
||||
{
|
||||
return split_huge_page_to_list(page, NULL);
|
||||
}
|
||||
void deferred_split_huge_page(struct page *page);
|
||||
void deferred_split_folio(struct folio *folio);
|
||||
|
||||
void __split_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd,
|
||||
unsigned long address, bool freeze, struct folio *folio);
|
||||
@ -340,7 +340,7 @@ static inline int split_huge_page(struct page *page)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline void deferred_split_huge_page(struct page *page) {}
|
||||
static inline void deferred_split_folio(struct folio *folio) {}
|
||||
#define split_huge_pmd(__vma, __pmd, __address) \
|
||||
do { } while (0)
|
||||
|
||||
|
@ -2815,9 +2815,8 @@ void free_transhuge_page(struct page *page)
|
||||
free_compound_page(page);
|
||||
}
|
||||
|
||||
void deferred_split_huge_page(struct page *page)
|
||||
void deferred_split_folio(struct folio *folio)
|
||||
{
|
||||
struct folio *folio = page_folio(page);
|
||||
struct deferred_split *ds_queue = get_deferred_split_queue(folio);
|
||||
#ifdef CONFIG_MEMCG
|
||||
struct mem_cgroup *memcg = folio_memcg(folio);
|
||||
|
Loading…
Reference in New Issue
Block a user