mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-12-29 09:16:33 +00:00
mm, slab: reintroduce rcu_barrier() into kmem_cache_destroy()
There used to be a rcu_barrier() for SLAB_TYPESAFE_BY_RCU caches in kmem_cache_destroy() until commit657dc2f972
("slab: remove synchronous rcu_barrier() call in memcg cache release path") moved it to an asynchronous work that finishes the destroying of such caches. The motivation for that commit was the MEMCG_KMEM integration that at the time created and removed clones of the global slab caches together with their cgroups, and blocking cgroups removal was unwelcome. The implementation later changed to per-object memcg tracking using a single cache, so there should be no more need for a fast non-blocking kmem_cache_destroy(), which is typically only done when a module is unloaded etc. Going back to synchronous barrier has the following advantages: - simpler implementation - it's easier to test the result of kmem_cache_destroy() in a kunit test Thus effectively revert commit657dc2f972
. It is not a 1:1 revert as the code has changed since. The main part is that kmem_cache_release(s) is always called from kmem_cache_destroy(), but for SLAB_TYPESAFE_BY_RCU caches there's a rcu_barrier() first. Suggested-by: Mateusz Guzik <mjguzik@gmail.com> Reviewed-by: Jann Horn <jannh@google.com> Signed-off-by: Vlastimil Babka <vbabka@suse.cz>
This commit is contained in:
parent
f77d0cda4a
commit
2eb14c1c27
@ -40,11 +40,6 @@ LIST_HEAD(slab_caches);
|
||||
DEFINE_MUTEX(slab_mutex);
|
||||
struct kmem_cache *kmem_cache;
|
||||
|
||||
static LIST_HEAD(slab_caches_to_rcu_destroy);
|
||||
static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work);
|
||||
static DECLARE_WORK(slab_caches_to_rcu_destroy_work,
|
||||
slab_caches_to_rcu_destroy_workfn);
|
||||
|
||||
/*
|
||||
* Set of flags that will prevent slab merging
|
||||
*/
|
||||
@ -499,33 +494,6 @@ static void kmem_cache_release(struct kmem_cache *s)
|
||||
slab_kmem_cache_release(s);
|
||||
}
|
||||
|
||||
static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work)
|
||||
{
|
||||
LIST_HEAD(to_destroy);
|
||||
struct kmem_cache *s, *s2;
|
||||
|
||||
/*
|
||||
* On destruction, SLAB_TYPESAFE_BY_RCU kmem_caches are put on the
|
||||
* @slab_caches_to_rcu_destroy list. The slab pages are freed
|
||||
* through RCU and the associated kmem_cache are dereferenced
|
||||
* while freeing the pages, so the kmem_caches should be freed only
|
||||
* after the pending RCU operations are finished. As rcu_barrier()
|
||||
* is a pretty slow operation, we batch all pending destructions
|
||||
* asynchronously.
|
||||
*/
|
||||
mutex_lock(&slab_mutex);
|
||||
list_splice_init(&slab_caches_to_rcu_destroy, &to_destroy);
|
||||
mutex_unlock(&slab_mutex);
|
||||
|
||||
if (list_empty(&to_destroy))
|
||||
return;
|
||||
|
||||
rcu_barrier();
|
||||
|
||||
list_for_each_entry_safe(s, s2, &to_destroy, list)
|
||||
kmem_cache_release(s);
|
||||
}
|
||||
|
||||
void slab_kmem_cache_release(struct kmem_cache *s)
|
||||
{
|
||||
__kmem_cache_release(s);
|
||||
@ -535,7 +503,6 @@ void slab_kmem_cache_release(struct kmem_cache *s)
|
||||
|
||||
void kmem_cache_destroy(struct kmem_cache *s)
|
||||
{
|
||||
bool rcu_set;
|
||||
int err;
|
||||
|
||||
if (unlikely(!s) || !kasan_check_byte(s))
|
||||
@ -551,8 +518,6 @@ void kmem_cache_destroy(struct kmem_cache *s)
|
||||
return;
|
||||
}
|
||||
|
||||
rcu_set = s->flags & SLAB_TYPESAFE_BY_RCU;
|
||||
|
||||
/* free asan quarantined objects */
|
||||
kasan_cache_shutdown(s);
|
||||
|
||||
@ -572,14 +537,10 @@ void kmem_cache_destroy(struct kmem_cache *s)
|
||||
if (err)
|
||||
return;
|
||||
|
||||
if (rcu_set) {
|
||||
mutex_lock(&slab_mutex);
|
||||
list_add_tail(&s->list, &slab_caches_to_rcu_destroy);
|
||||
schedule_work(&slab_caches_to_rcu_destroy_work);
|
||||
mutex_unlock(&slab_mutex);
|
||||
} else {
|
||||
kmem_cache_release(s);
|
||||
}
|
||||
if (s->flags & SLAB_TYPESAFE_BY_RCU)
|
||||
rcu_barrier();
|
||||
|
||||
kmem_cache_release(s);
|
||||
}
|
||||
EXPORT_SYMBOL(kmem_cache_destroy);
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user