mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2025-01-01 10:45:49 +00:00
176 lines
4.0 KiB
C
176 lines
4.0 KiB
C
|
// SPDX-License-Identifier: GPL-2.0-only
|
||
|
/* Copyright (c) 2024 Google */
|
||
|
#include <linux/bpf.h>
|
||
|
#include <linux/btf_ids.h>
|
||
|
#include <linux/slab.h>
|
||
|
#include <linux/kernel.h>
|
||
|
#include <linux/seq_file.h>
|
||
|
|
||
|
#include "../../mm/slab.h" /* kmem_cache, slab_caches and slab_mutex */
|
||
|
|
||
|
struct bpf_iter__kmem_cache {
|
||
|
__bpf_md_ptr(struct bpf_iter_meta *, meta);
|
||
|
__bpf_md_ptr(struct kmem_cache *, s);
|
||
|
};
|
||
|
|
||
|
static void *kmem_cache_iter_seq_start(struct seq_file *seq, loff_t *pos)
|
||
|
{
|
||
|
loff_t cnt = 0;
|
||
|
bool found = false;
|
||
|
struct kmem_cache *s;
|
||
|
|
||
|
mutex_lock(&slab_mutex);
|
||
|
|
||
|
/* Find an entry at the given position in the slab_caches list instead
|
||
|
* of keeping a reference (of the last visited entry, if any) out of
|
||
|
* slab_mutex. It might miss something if one is deleted in the middle
|
||
|
* while it releases the lock. But it should be rare and there's not
|
||
|
* much we can do about it.
|
||
|
*/
|
||
|
list_for_each_entry(s, &slab_caches, list) {
|
||
|
if (cnt == *pos) {
|
||
|
/* Make sure this entry remains in the list by getting
|
||
|
* a new reference count. Note that boot_cache entries
|
||
|
* have a negative refcount, so don't touch them.
|
||
|
*/
|
||
|
if (s->refcount > 0)
|
||
|
s->refcount++;
|
||
|
found = true;
|
||
|
break;
|
||
|
}
|
||
|
cnt++;
|
||
|
}
|
||
|
mutex_unlock(&slab_mutex);
|
||
|
|
||
|
if (!found)
|
||
|
return NULL;
|
||
|
|
||
|
return s;
|
||
|
}
|
||
|
|
||
|
static void kmem_cache_iter_seq_stop(struct seq_file *seq, void *v)
|
||
|
{
|
||
|
struct bpf_iter_meta meta;
|
||
|
struct bpf_iter__kmem_cache ctx = {
|
||
|
.meta = &meta,
|
||
|
.s = v,
|
||
|
};
|
||
|
struct bpf_prog *prog;
|
||
|
bool destroy = false;
|
||
|
|
||
|
meta.seq = seq;
|
||
|
prog = bpf_iter_get_info(&meta, true);
|
||
|
if (prog && !ctx.s)
|
||
|
bpf_iter_run_prog(prog, &ctx);
|
||
|
|
||
|
if (ctx.s == NULL)
|
||
|
return;
|
||
|
|
||
|
mutex_lock(&slab_mutex);
|
||
|
|
||
|
/* Skip kmem_cache_destroy() for active entries */
|
||
|
if (ctx.s->refcount > 1)
|
||
|
ctx.s->refcount--;
|
||
|
else if (ctx.s->refcount == 1)
|
||
|
destroy = true;
|
||
|
|
||
|
mutex_unlock(&slab_mutex);
|
||
|
|
||
|
if (destroy)
|
||
|
kmem_cache_destroy(ctx.s);
|
||
|
}
|
||
|
|
||
|
static void *kmem_cache_iter_seq_next(struct seq_file *seq, void *v, loff_t *pos)
|
||
|
{
|
||
|
struct kmem_cache *s = v;
|
||
|
struct kmem_cache *next = NULL;
|
||
|
bool destroy = false;
|
||
|
|
||
|
++*pos;
|
||
|
|
||
|
mutex_lock(&slab_mutex);
|
||
|
|
||
|
if (list_last_entry(&slab_caches, struct kmem_cache, list) != s) {
|
||
|
next = list_next_entry(s, list);
|
||
|
|
||
|
WARN_ON_ONCE(next->refcount == 0);
|
||
|
|
||
|
/* boot_caches have negative refcount, don't touch them */
|
||
|
if (next->refcount > 0)
|
||
|
next->refcount++;
|
||
|
}
|
||
|
|
||
|
/* Skip kmem_cache_destroy() for active entries */
|
||
|
if (s->refcount > 1)
|
||
|
s->refcount--;
|
||
|
else if (s->refcount == 1)
|
||
|
destroy = true;
|
||
|
|
||
|
mutex_unlock(&slab_mutex);
|
||
|
|
||
|
if (destroy)
|
||
|
kmem_cache_destroy(s);
|
||
|
|
||
|
return next;
|
||
|
}
|
||
|
|
||
|
static int kmem_cache_iter_seq_show(struct seq_file *seq, void *v)
|
||
|
{
|
||
|
struct bpf_iter_meta meta;
|
||
|
struct bpf_iter__kmem_cache ctx = {
|
||
|
.meta = &meta,
|
||
|
.s = v,
|
||
|
};
|
||
|
struct bpf_prog *prog;
|
||
|
int ret = 0;
|
||
|
|
||
|
meta.seq = seq;
|
||
|
prog = bpf_iter_get_info(&meta, false);
|
||
|
if (prog)
|
||
|
ret = bpf_iter_run_prog(prog, &ctx);
|
||
|
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
static const struct seq_operations kmem_cache_iter_seq_ops = {
|
||
|
.start = kmem_cache_iter_seq_start,
|
||
|
.next = kmem_cache_iter_seq_next,
|
||
|
.stop = kmem_cache_iter_seq_stop,
|
||
|
.show = kmem_cache_iter_seq_show,
|
||
|
};
|
||
|
|
||
|
BTF_ID_LIST_GLOBAL_SINGLE(bpf_kmem_cache_btf_id, struct, kmem_cache)
|
||
|
|
||
|
static const struct bpf_iter_seq_info kmem_cache_iter_seq_info = {
|
||
|
.seq_ops = &kmem_cache_iter_seq_ops,
|
||
|
};
|
||
|
|
||
|
static void bpf_iter_kmem_cache_show_fdinfo(const struct bpf_iter_aux_info *aux,
|
||
|
struct seq_file *seq)
|
||
|
{
|
||
|
seq_puts(seq, "kmem_cache iter\n");
|
||
|
}
|
||
|
|
||
|
DEFINE_BPF_ITER_FUNC(kmem_cache, struct bpf_iter_meta *meta,
|
||
|
struct kmem_cache *s)
|
||
|
|
||
|
static struct bpf_iter_reg bpf_kmem_cache_reg_info = {
|
||
|
.target = "kmem_cache",
|
||
|
.feature = BPF_ITER_RESCHED,
|
||
|
.show_fdinfo = bpf_iter_kmem_cache_show_fdinfo,
|
||
|
.ctx_arg_info_size = 1,
|
||
|
.ctx_arg_info = {
|
||
|
{ offsetof(struct bpf_iter__kmem_cache, s),
|
||
|
PTR_TO_BTF_ID_OR_NULL | PTR_TRUSTED },
|
||
|
},
|
||
|
.seq_info = &kmem_cache_iter_seq_info,
|
||
|
};
|
||
|
|
||
|
static int __init bpf_kmem_cache_iter_init(void)
|
||
|
{
|
||
|
bpf_kmem_cache_reg_info.ctx_arg_info[0].btf_id = bpf_kmem_cache_btf_id[0];
|
||
|
return bpf_iter_reg_target(&bpf_kmem_cache_reg_info);
|
||
|
}
|
||
|
|
||
|
late_initcall(bpf_kmem_cache_iter_init);
|