mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2024-12-29 01:03:32 +00:00
2c321f3f70
Main goal of memory allocation profiling patchset is to provide accounting that is cheap enough to run in production. To achieve that we inject counters using codetags at the allocation call sites to account every time allocation is made. This injection allows us to perform accounting efficiently because injected counters are immediately available as opposed to the alternative methods, such as using _RET_IP_, which would require counter lookup and appropriate locking that makes accounting much more expensive. This method requires all allocation functions to inject separate counters at their call sites so that their callers can be individually accounted. Counter injection is implemented by allocation hooks which should wrap all allocation functions. Inlined functions which perform allocations but do not use allocation hooks are directly charged for the allocations they perform. In most cases these functions are just specialized allocation wrappers used from multiple places to allocate objects of a specific type. It would be more useful to do the accounting at their call sites instead. Instrument these helpers to do accounting at the call site. Simple inlined allocation wrappers are converted directly into macros. More complex allocators or allocators with documentation are converted into _noprof versions and allocation hooks are added. This allows memory allocation profiling mechanism to charge allocations to the callers of these functions. Link: https://lkml.kernel.org/r/20240415020731.1152108-1-surenb@google.com Signed-off-by: Suren Baghdasaryan <surenb@google.com> Acked-by: Jan Kara <jack@suse.cz> [jbd2] Cc: Anna Schumaker <anna@kernel.org> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Benjamin Tissoires <benjamin.tissoires@redhat.com> Cc: Christoph Lameter <cl@linux.com> Cc: David Rientjes <rientjes@google.com> Cc: David S. Miller <davem@davemloft.net> Cc: Dennis Zhou <dennis@kernel.org> Cc: Eric Dumazet <edumazet@google.com> Cc: Herbert Xu <herbert@gondor.apana.org.au> Cc: Jakub Kicinski <kuba@kernel.org> Cc: Jakub Sitnicki <jakub@cloudflare.com> Cc: Jiri Kosina <jikos@kernel.org> Cc: Joerg Roedel <joro@8bytes.org> Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com> Cc: Kent Overstreet <kent.overstreet@linux.dev> Cc: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: Paolo Abeni <pabeni@redhat.com> Cc: Pekka Enberg <penberg@kernel.org> Cc: Tejun Heo <tj@kernel.org> Cc: Theodore Ts'o <tytso@mit.edu> Cc: Trond Myklebust <trond.myklebust@hammerspace.com> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Will Deacon <will@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
228 lines
6.0 KiB
C
228 lines
6.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __ASM_GENERIC_PGALLOC_H
|
|
#define __ASM_GENERIC_PGALLOC_H
|
|
|
|
#ifdef CONFIG_MMU
|
|
|
|
#define GFP_PGTABLE_KERNEL (GFP_KERNEL | __GFP_ZERO)
|
|
#define GFP_PGTABLE_USER (GFP_PGTABLE_KERNEL | __GFP_ACCOUNT)
|
|
|
|
/**
|
|
* __pte_alloc_one_kernel - allocate memory for a PTE-level kernel page table
|
|
* @mm: the mm_struct of the current context
|
|
*
|
|
* This function is intended for architectures that need
|
|
* anything beyond simple page allocation.
|
|
*
|
|
* Return: pointer to the allocated memory or %NULL on error
|
|
*/
|
|
static inline pte_t *__pte_alloc_one_kernel_noprof(struct mm_struct *mm)
|
|
{
|
|
struct ptdesc *ptdesc = pagetable_alloc_noprof(GFP_PGTABLE_KERNEL &
|
|
~__GFP_HIGHMEM, 0);
|
|
|
|
if (!ptdesc)
|
|
return NULL;
|
|
return ptdesc_address(ptdesc);
|
|
}
|
|
#define __pte_alloc_one_kernel(...) alloc_hooks(__pte_alloc_one_kernel_noprof(__VA_ARGS__))
|
|
|
|
#ifndef __HAVE_ARCH_PTE_ALLOC_ONE_KERNEL
|
|
/**
|
|
* pte_alloc_one_kernel - allocate memory for a PTE-level kernel page table
|
|
* @mm: the mm_struct of the current context
|
|
*
|
|
* Return: pointer to the allocated memory or %NULL on error
|
|
*/
|
|
static inline pte_t *pte_alloc_one_kernel_noprof(struct mm_struct *mm)
|
|
{
|
|
return __pte_alloc_one_kernel_noprof(mm);
|
|
}
|
|
#define pte_alloc_one_kernel(...) alloc_hooks(pte_alloc_one_kernel_noprof(__VA_ARGS__))
|
|
#endif
|
|
|
|
/**
|
|
* pte_free_kernel - free PTE-level kernel page table memory
|
|
* @mm: the mm_struct of the current context
|
|
* @pte: pointer to the memory containing the page table
|
|
*/
|
|
static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte)
|
|
{
|
|
pagetable_free(virt_to_ptdesc(pte));
|
|
}
|
|
|
|
/**
|
|
* __pte_alloc_one - allocate memory for a PTE-level user page table
|
|
* @mm: the mm_struct of the current context
|
|
* @gfp: GFP flags to use for the allocation
|
|
*
|
|
* Allocate memory for a page table and ptdesc and runs pagetable_pte_ctor().
|
|
*
|
|
* This function is intended for architectures that need
|
|
* anything beyond simple page allocation or must have custom GFP flags.
|
|
*
|
|
* Return: `struct page` referencing the ptdesc or %NULL on error
|
|
*/
|
|
static inline pgtable_t __pte_alloc_one_noprof(struct mm_struct *mm, gfp_t gfp)
|
|
{
|
|
struct ptdesc *ptdesc;
|
|
|
|
ptdesc = pagetable_alloc_noprof(gfp, 0);
|
|
if (!ptdesc)
|
|
return NULL;
|
|
if (!pagetable_pte_ctor(ptdesc)) {
|
|
pagetable_free(ptdesc);
|
|
return NULL;
|
|
}
|
|
|
|
return ptdesc_page(ptdesc);
|
|
}
|
|
#define __pte_alloc_one(...) alloc_hooks(__pte_alloc_one_noprof(__VA_ARGS__))
|
|
|
|
#ifndef __HAVE_ARCH_PTE_ALLOC_ONE
|
|
/**
|
|
* pte_alloc_one - allocate a page for PTE-level user page table
|
|
* @mm: the mm_struct of the current context
|
|
*
|
|
* Allocate memory for a page table and ptdesc and runs pagetable_pte_ctor().
|
|
*
|
|
* Return: `struct page` referencing the ptdesc or %NULL on error
|
|
*/
|
|
static inline pgtable_t pte_alloc_one_noprof(struct mm_struct *mm)
|
|
{
|
|
return __pte_alloc_one_noprof(mm, GFP_PGTABLE_USER);
|
|
}
|
|
#define pte_alloc_one(...) alloc_hooks(pte_alloc_one_noprof(__VA_ARGS__))
|
|
#endif
|
|
|
|
/*
|
|
* Should really implement gc for free page table pages. This could be
|
|
* done with a reference count in struct page.
|
|
*/
|
|
|
|
/**
|
|
* pte_free - free PTE-level user page table memory
|
|
* @mm: the mm_struct of the current context
|
|
* @pte_page: the `struct page` referencing the ptdesc
|
|
*/
|
|
static inline void pte_free(struct mm_struct *mm, struct page *pte_page)
|
|
{
|
|
struct ptdesc *ptdesc = page_ptdesc(pte_page);
|
|
|
|
pagetable_pte_dtor(ptdesc);
|
|
pagetable_free(ptdesc);
|
|
}
|
|
|
|
|
|
#if CONFIG_PGTABLE_LEVELS > 2
|
|
|
|
#ifndef __HAVE_ARCH_PMD_ALLOC_ONE
|
|
/**
|
|
* pmd_alloc_one - allocate memory for a PMD-level page table
|
|
* @mm: the mm_struct of the current context
|
|
*
|
|
* Allocate memory for a page table and ptdesc and runs pagetable_pmd_ctor().
|
|
*
|
|
* Allocations use %GFP_PGTABLE_USER in user context and
|
|
* %GFP_PGTABLE_KERNEL in kernel context.
|
|
*
|
|
* Return: pointer to the allocated memory or %NULL on error
|
|
*/
|
|
static inline pmd_t *pmd_alloc_one_noprof(struct mm_struct *mm, unsigned long addr)
|
|
{
|
|
struct ptdesc *ptdesc;
|
|
gfp_t gfp = GFP_PGTABLE_USER;
|
|
|
|
if (mm == &init_mm)
|
|
gfp = GFP_PGTABLE_KERNEL;
|
|
ptdesc = pagetable_alloc_noprof(gfp, 0);
|
|
if (!ptdesc)
|
|
return NULL;
|
|
if (!pagetable_pmd_ctor(ptdesc)) {
|
|
pagetable_free(ptdesc);
|
|
return NULL;
|
|
}
|
|
return ptdesc_address(ptdesc);
|
|
}
|
|
#define pmd_alloc_one(...) alloc_hooks(pmd_alloc_one_noprof(__VA_ARGS__))
|
|
#endif
|
|
|
|
#ifndef __HAVE_ARCH_PMD_FREE
|
|
static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd)
|
|
{
|
|
struct ptdesc *ptdesc = virt_to_ptdesc(pmd);
|
|
|
|
BUG_ON((unsigned long)pmd & (PAGE_SIZE-1));
|
|
pagetable_pmd_dtor(ptdesc);
|
|
pagetable_free(ptdesc);
|
|
}
|
|
#endif
|
|
|
|
#endif /* CONFIG_PGTABLE_LEVELS > 2 */
|
|
|
|
#if CONFIG_PGTABLE_LEVELS > 3
|
|
|
|
static inline pud_t *__pud_alloc_one_noprof(struct mm_struct *mm, unsigned long addr)
|
|
{
|
|
gfp_t gfp = GFP_PGTABLE_USER;
|
|
struct ptdesc *ptdesc;
|
|
|
|
if (mm == &init_mm)
|
|
gfp = GFP_PGTABLE_KERNEL;
|
|
gfp &= ~__GFP_HIGHMEM;
|
|
|
|
ptdesc = pagetable_alloc_noprof(gfp, 0);
|
|
if (!ptdesc)
|
|
return NULL;
|
|
|
|
pagetable_pud_ctor(ptdesc);
|
|
return ptdesc_address(ptdesc);
|
|
}
|
|
#define __pud_alloc_one(...) alloc_hooks(__pud_alloc_one_noprof(__VA_ARGS__))
|
|
|
|
#ifndef __HAVE_ARCH_PUD_ALLOC_ONE
|
|
/**
|
|
* pud_alloc_one - allocate memory for a PUD-level page table
|
|
* @mm: the mm_struct of the current context
|
|
*
|
|
* Allocate memory for a page table using %GFP_PGTABLE_USER for user context
|
|
* and %GFP_PGTABLE_KERNEL for kernel context.
|
|
*
|
|
* Return: pointer to the allocated memory or %NULL on error
|
|
*/
|
|
static inline pud_t *pud_alloc_one_noprof(struct mm_struct *mm, unsigned long addr)
|
|
{
|
|
return __pud_alloc_one_noprof(mm, addr);
|
|
}
|
|
#define pud_alloc_one(...) alloc_hooks(pud_alloc_one_noprof(__VA_ARGS__))
|
|
#endif
|
|
|
|
static inline void __pud_free(struct mm_struct *mm, pud_t *pud)
|
|
{
|
|
struct ptdesc *ptdesc = virt_to_ptdesc(pud);
|
|
|
|
BUG_ON((unsigned long)pud & (PAGE_SIZE-1));
|
|
pagetable_pud_dtor(ptdesc);
|
|
pagetable_free(ptdesc);
|
|
}
|
|
|
|
#ifndef __HAVE_ARCH_PUD_FREE
|
|
static inline void pud_free(struct mm_struct *mm, pud_t *pud)
|
|
{
|
|
__pud_free(mm, pud);
|
|
}
|
|
#endif
|
|
|
|
#endif /* CONFIG_PGTABLE_LEVELS > 3 */
|
|
|
|
#ifndef __HAVE_ARCH_PGD_FREE
|
|
static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd)
|
|
{
|
|
pagetable_free(virt_to_ptdesc(pgd));
|
|
}
|
|
#endif
|
|
|
|
#endif /* CONFIG_MMU */
|
|
|
|
#endif /* __ASM_GENERIC_PGALLOC_H */
|