xfs: use per-mount cpumask to track nonempty percpu inodegc lists

Directly track which CPUs have contributed to the inodegc percpu lists
instead of trusting the cpu online mask.  This eliminates a theoretical
problem where the inodegc flush functions might fail to flush a CPU's
inodes if that CPU happened to be dying at exactly the same time.  Most
likely nobody's noticed this because the CPU dead hook moves the percpu
inodegc list to another CPU and schedules that worker immediately.  But
it's quite possible that this is a subtle race leading to UAF if the
inodegc flush were part of an unmount.

Further benefits: This reduces the overhead of the inodegc flush code
slightly by allowing us to ignore CPUs that have empty lists.  Better
yet, it reduces our dependence on the cpu online masks, which have been
the cause of confusion and drama lately.

Fixes: ab23a77687 ("xfs: per-cpu deferred inode inactivation queues")
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
This commit is contained in:
Darrick J. Wong 2023-09-11 08:39:03 -07:00
parent ecd49f7a36
commit 62334fab47
4 changed files with 33 additions and 56 deletions

View File

@ -443,7 +443,7 @@ xfs_inodegc_queue_all(
int cpu; int cpu;
bool ret = false; bool ret = false;
for_each_online_cpu(cpu) { for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
gc = per_cpu_ptr(mp->m_inodegc, cpu); gc = per_cpu_ptr(mp->m_inodegc, cpu);
if (!llist_empty(&gc->list)) { if (!llist_empty(&gc->list)) {
mod_delayed_work_on(cpu, mp->m_inodegc_wq, &gc->work, 0); mod_delayed_work_on(cpu, mp->m_inodegc_wq, &gc->work, 0);
@ -463,7 +463,7 @@ xfs_inodegc_wait_all(
int error = 0; int error = 0;
flush_workqueue(mp->m_inodegc_wq); flush_workqueue(mp->m_inodegc_wq);
for_each_online_cpu(cpu) { for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
struct xfs_inodegc *gc; struct xfs_inodegc *gc;
gc = per_cpu_ptr(mp->m_inodegc, cpu); gc = per_cpu_ptr(mp->m_inodegc, cpu);
@ -1845,9 +1845,17 @@ xfs_inodegc_worker(
struct xfs_inodegc, work); struct xfs_inodegc, work);
struct llist_node *node = llist_del_all(&gc->list); struct llist_node *node = llist_del_all(&gc->list);
struct xfs_inode *ip, *n; struct xfs_inode *ip, *n;
struct xfs_mount *mp = gc->mp;
unsigned int nofs_flag; unsigned int nofs_flag;
ASSERT(gc->cpu == smp_processor_id()); /*
* Clear the cpu mask bit and ensure that we have seen the latest
* update of the gc structure associated with this CPU. This matches
* with the release semantics used when setting the cpumask bit in
* xfs_inodegc_queue.
*/
cpumask_clear_cpu(gc->cpu, &mp->m_inodegc_cpumask);
smp_mb__after_atomic();
WRITE_ONCE(gc->items, 0); WRITE_ONCE(gc->items, 0);
@ -1862,7 +1870,7 @@ xfs_inodegc_worker(
nofs_flag = memalloc_nofs_save(); nofs_flag = memalloc_nofs_save();
ip = llist_entry(node, struct xfs_inode, i_gclist); ip = llist_entry(node, struct xfs_inode, i_gclist);
trace_xfs_inodegc_worker(ip->i_mount, READ_ONCE(gc->shrinker_hits)); trace_xfs_inodegc_worker(mp, READ_ONCE(gc->shrinker_hits));
WRITE_ONCE(gc->shrinker_hits, 0); WRITE_ONCE(gc->shrinker_hits, 0);
llist_for_each_entry_safe(ip, n, node, i_gclist) { llist_for_each_entry_safe(ip, n, node, i_gclist) {
@ -2057,6 +2065,7 @@ xfs_inodegc_queue(
struct xfs_inodegc *gc; struct xfs_inodegc *gc;
int items; int items;
unsigned int shrinker_hits; unsigned int shrinker_hits;
unsigned int cpu_nr;
unsigned long queue_delay = 1; unsigned long queue_delay = 1;
trace_xfs_inode_set_need_inactive(ip); trace_xfs_inode_set_need_inactive(ip);
@ -2064,18 +2073,28 @@ xfs_inodegc_queue(
ip->i_flags |= XFS_NEED_INACTIVE; ip->i_flags |= XFS_NEED_INACTIVE;
spin_unlock(&ip->i_flags_lock); spin_unlock(&ip->i_flags_lock);
gc = get_cpu_ptr(mp->m_inodegc); cpu_nr = get_cpu();
gc = this_cpu_ptr(mp->m_inodegc);
llist_add(&ip->i_gclist, &gc->list); llist_add(&ip->i_gclist, &gc->list);
items = READ_ONCE(gc->items); items = READ_ONCE(gc->items);
WRITE_ONCE(gc->items, items + 1); WRITE_ONCE(gc->items, items + 1);
shrinker_hits = READ_ONCE(gc->shrinker_hits); shrinker_hits = READ_ONCE(gc->shrinker_hits);
/*
* Ensure the list add is always seen by anyone who finds the cpumask
* bit set. This effectively gives the cpumask bit set operation
* release ordering semantics.
*/
smp_mb__before_atomic();
if (!cpumask_test_cpu(cpu_nr, &mp->m_inodegc_cpumask))
cpumask_test_and_set_cpu(cpu_nr, &mp->m_inodegc_cpumask);
/* /*
* We queue the work while holding the current CPU so that the work * We queue the work while holding the current CPU so that the work
* is scheduled to run on this CPU. * is scheduled to run on this CPU.
*/ */
if (!xfs_is_inodegc_enabled(mp)) { if (!xfs_is_inodegc_enabled(mp)) {
put_cpu_ptr(gc); put_cpu();
return; return;
} }
@ -2085,7 +2104,7 @@ xfs_inodegc_queue(
trace_xfs_inodegc_queue(mp, __return_address); trace_xfs_inodegc_queue(mp, __return_address);
mod_delayed_work_on(current_cpu(), mp->m_inodegc_wq, &gc->work, mod_delayed_work_on(current_cpu(), mp->m_inodegc_wq, &gc->work,
queue_delay); queue_delay);
put_cpu_ptr(gc); put_cpu();
if (xfs_inodegc_want_flush_work(ip, items, shrinker_hits)) { if (xfs_inodegc_want_flush_work(ip, items, shrinker_hits)) {
trace_xfs_inodegc_throttle(mp, __return_address); trace_xfs_inodegc_throttle(mp, __return_address);
@ -2093,47 +2112,6 @@ xfs_inodegc_queue(
} }
} }
/*
* Fold the dead CPU inodegc queue into the current CPUs queue.
*/
void
xfs_inodegc_cpu_dead(
struct xfs_mount *mp,
unsigned int dead_cpu)
{
struct xfs_inodegc *dead_gc, *gc;
struct llist_node *first, *last;
unsigned int count = 0;
dead_gc = per_cpu_ptr(mp->m_inodegc, dead_cpu);
cancel_delayed_work_sync(&dead_gc->work);
if (llist_empty(&dead_gc->list))
return;
first = dead_gc->list.first;
last = first;
while (last->next) {
last = last->next;
count++;
}
dead_gc->list.first = NULL;
dead_gc->items = 0;
/* Add pending work to current CPU */
gc = get_cpu_ptr(mp->m_inodegc);
llist_add_batch(first, last, &gc->list);
count += READ_ONCE(gc->items);
WRITE_ONCE(gc->items, count);
if (xfs_is_inodegc_enabled(mp)) {
trace_xfs_inodegc_queue(mp, __return_address);
mod_delayed_work_on(current_cpu(), mp->m_inodegc_wq, &gc->work,
0);
}
put_cpu_ptr(gc);
}
/* /*
* We set the inode flag atomically with the radix tree tag. Once we get tag * We set the inode flag atomically with the radix tree tag. Once we get tag
* lookups on the radix tree, this inode flag can go away. * lookups on the radix tree, this inode flag can go away.
@ -2195,7 +2173,7 @@ xfs_inodegc_shrinker_count(
if (!xfs_is_inodegc_enabled(mp)) if (!xfs_is_inodegc_enabled(mp))
return 0; return 0;
for_each_online_cpu(cpu) { for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
gc = per_cpu_ptr(mp->m_inodegc, cpu); gc = per_cpu_ptr(mp->m_inodegc, cpu);
if (!llist_empty(&gc->list)) if (!llist_empty(&gc->list))
return XFS_INODEGC_SHRINKER_COUNT; return XFS_INODEGC_SHRINKER_COUNT;
@ -2220,7 +2198,7 @@ xfs_inodegc_shrinker_scan(
trace_xfs_inodegc_shrinker_scan(mp, sc, __return_address); trace_xfs_inodegc_shrinker_scan(mp, sc, __return_address);
for_each_online_cpu(cpu) { for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
gc = per_cpu_ptr(mp->m_inodegc, cpu); gc = per_cpu_ptr(mp->m_inodegc, cpu);
if (!llist_empty(&gc->list)) { if (!llist_empty(&gc->list)) {
unsigned int h = READ_ONCE(gc->shrinker_hits); unsigned int h = READ_ONCE(gc->shrinker_hits);

View File

@ -79,7 +79,6 @@ void xfs_inodegc_push(struct xfs_mount *mp);
int xfs_inodegc_flush(struct xfs_mount *mp); int xfs_inodegc_flush(struct xfs_mount *mp);
void xfs_inodegc_stop(struct xfs_mount *mp); void xfs_inodegc_stop(struct xfs_mount *mp);
void xfs_inodegc_start(struct xfs_mount *mp); void xfs_inodegc_start(struct xfs_mount *mp);
void xfs_inodegc_cpu_dead(struct xfs_mount *mp, unsigned int cpu);
int xfs_inodegc_register_shrinker(struct xfs_mount *mp); int xfs_inodegc_register_shrinker(struct xfs_mount *mp);
#endif #endif

View File

@ -60,6 +60,7 @@ struct xfs_error_cfg {
* Per-cpu deferred inode inactivation GC lists. * Per-cpu deferred inode inactivation GC lists.
*/ */
struct xfs_inodegc { struct xfs_inodegc {
struct xfs_mount *mp;
struct llist_head list; struct llist_head list;
struct delayed_work work; struct delayed_work work;
int error; int error;
@ -67,9 +68,7 @@ struct xfs_inodegc {
/* approximate count of inodes in the list */ /* approximate count of inodes in the list */
unsigned int items; unsigned int items;
unsigned int shrinker_hits; unsigned int shrinker_hits;
#if defined(DEBUG) || defined(XFS_WARN)
unsigned int cpu; unsigned int cpu;
#endif
}; };
/* /*
@ -249,6 +248,9 @@ typedef struct xfs_mount {
unsigned int *m_errortag; unsigned int *m_errortag;
struct xfs_kobj m_errortag_kobj; struct xfs_kobj m_errortag_kobj;
#endif #endif
/* cpus that have inodes queued for inactivation */
struct cpumask m_inodegc_cpumask;
} xfs_mount_t; } xfs_mount_t;
#define M_IGEO(mp) (&(mp)->m_ino_geo) #define M_IGEO(mp) (&(mp)->m_ino_geo)

View File

@ -1135,9 +1135,8 @@ xfs_inodegc_init_percpu(
for_each_possible_cpu(cpu) { for_each_possible_cpu(cpu) {
gc = per_cpu_ptr(mp->m_inodegc, cpu); gc = per_cpu_ptr(mp->m_inodegc, cpu);
#if defined(DEBUG) || defined(XFS_WARN)
gc->cpu = cpu; gc->cpu = cpu;
#endif gc->mp = mp;
init_llist_head(&gc->list); init_llist_head(&gc->list);
gc->items = 0; gc->items = 0;
gc->error = 0; gc->error = 0;
@ -2336,7 +2335,6 @@ xfs_cpu_dead(
spin_lock(&xfs_mount_list_lock); spin_lock(&xfs_mount_list_lock);
list_for_each_entry_safe(mp, n, &xfs_mount_list, m_mount_list) { list_for_each_entry_safe(mp, n, &xfs_mount_list, m_mount_list) {
spin_unlock(&xfs_mount_list_lock); spin_unlock(&xfs_mount_list_lock);
xfs_inodegc_cpu_dead(mp, cpu);
spin_lock(&xfs_mount_list_lock); spin_lock(&xfs_mount_list_lock);
} }
spin_unlock(&xfs_mount_list_lock); spin_unlock(&xfs_mount_list_lock);