mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-29 09:12:07 +00:00
xdp: Add proper __rcu annotations to redirect map entries
XDP_REDIRECT works by a three-step process: the bpf_redirect() and bpf_redirect_map() helpers will lookup the target of the redirect and store it (along with some other metadata) in a per-CPU struct bpf_redirect_info. Next, when the program returns the XDP_REDIRECT return code, the driver will call xdp_do_redirect() which will use the information thus stored to actually enqueue the frame into a bulk queue structure (that differs slightly by map type, but shares the same principle). Finally, before exiting its NAPI poll loop, the driver will call xdp_do_flush(), which will flush all the different bulk queues, thus completing the redirect. Pointers to the map entries will be kept around for this whole sequence of steps, protected by RCU. However, there is no top-level rcu_read_lock() in the core code; instead drivers add their own rcu_read_lock() around the XDP portions of the code, but somewhat inconsistently as Martin discovered[0]. However, things still work because everything happens inside a single NAPI poll sequence, which means it's between a pair of calls to local_bh_disable()/local_bh_enable(). So Paul suggested[1] that we could document this intention by using rcu_dereference_check() with rcu_read_lock_bh_held() as a second parameter, thus allowing sparse and lockdep to verify that everything is done correctly. This patch does just that: we add an __rcu annotation to the map entry pointers and remove the various comments explaining the NAPI poll assurance strewn through devmap.c in favour of a longer explanation in filter.c. The goal is to have one coherent documentation of the entire flow, and rely on the RCU annotations as a "standard" way of communicating the flow in the map code (which can additionally be understood by sparse and lockdep). The RCU annotation replacements result in a fairly straight-forward replacement where READ_ONCE() becomes rcu_dereference_check(), WRITE_ONCE() becomes rcu_assign_pointer() and xchg() and cmpxchg() gets wrapped in the proper constructs to cast the pointer back and forth between __rcu and __kernel address space (for the benefit of sparse). The one complication is that xskmap has a few constructions where double-pointers are passed back and forth; these simply all gain __rcu annotations, and only the final reference/dereference to the inner-most pointer gets changed. With this, everything can be run through sparse without eliciting complaints, and lockdep can verify correctness even without the use of rcu_read_lock() in the drivers. Subsequent patches will clean these up from the drivers. [0] https://lore.kernel.org/bpf/20210415173551.7ma4slcbqeyiba2r@kafai-mbp.dhcp.thefacebook.com/ [1] https://lore.kernel.org/bpf/20210419165837.GA975577@paulmck-ThinkPad-P17-Gen-1/ Signed-off-by: Toke Høiland-Jørgensen <toke@redhat.com> Signed-off-by: Daniel Borkmann <daniel@iogearbox.net> Link: https://lore.kernel.org/bpf/20210624160609.292325-6-toke@redhat.com
This commit is contained in:
parent
694cea395f
commit
782347b6bc
@ -763,11 +763,9 @@ DECLARE_BPF_DISPATCHER(xdp)
|
|||||||
static __always_inline u32 bpf_prog_run_xdp(const struct bpf_prog *prog,
|
static __always_inline u32 bpf_prog_run_xdp(const struct bpf_prog *prog,
|
||||||
struct xdp_buff *xdp)
|
struct xdp_buff *xdp)
|
||||||
{
|
{
|
||||||
/* Caller needs to hold rcu_read_lock() (!), otherwise program
|
/* Driver XDP hooks are invoked within a single NAPI poll cycle and thus
|
||||||
* can be released while still running, or map elements could be
|
* under local_bh_disable(), which provides the needed RCU protection
|
||||||
* freed early while still having concurrent users. XDP fastpath
|
* for accessing map entries.
|
||||||
* already takes rcu_read_lock() when fetching the program, so
|
|
||||||
* it's not necessary here anymore.
|
|
||||||
*/
|
*/
|
||||||
return __BPF_PROG_RUN(prog, xdp, BPF_DISPATCHER_FUNC(xdp));
|
return __BPF_PROG_RUN(prog, xdp, BPF_DISPATCHER_FUNC(xdp));
|
||||||
}
|
}
|
||||||
|
@ -37,7 +37,7 @@ struct xdp_umem {
|
|||||||
struct xsk_map {
|
struct xsk_map {
|
||||||
struct bpf_map map;
|
struct bpf_map map;
|
||||||
spinlock_t lock; /* Synchronize map updates */
|
spinlock_t lock; /* Synchronize map updates */
|
||||||
struct xdp_sock *xsk_map[];
|
struct xdp_sock __rcu *xsk_map[];
|
||||||
};
|
};
|
||||||
|
|
||||||
struct xdp_sock {
|
struct xdp_sock {
|
||||||
|
@ -74,7 +74,7 @@ struct bpf_cpu_map_entry {
|
|||||||
struct bpf_cpu_map {
|
struct bpf_cpu_map {
|
||||||
struct bpf_map map;
|
struct bpf_map map;
|
||||||
/* Below members specific for map type */
|
/* Below members specific for map type */
|
||||||
struct bpf_cpu_map_entry **cpu_map;
|
struct bpf_cpu_map_entry __rcu **cpu_map;
|
||||||
};
|
};
|
||||||
|
|
||||||
static DEFINE_PER_CPU(struct list_head, cpu_map_flush_list);
|
static DEFINE_PER_CPU(struct list_head, cpu_map_flush_list);
|
||||||
@ -469,7 +469,7 @@ static void __cpu_map_entry_replace(struct bpf_cpu_map *cmap,
|
|||||||
{
|
{
|
||||||
struct bpf_cpu_map_entry *old_rcpu;
|
struct bpf_cpu_map_entry *old_rcpu;
|
||||||
|
|
||||||
old_rcpu = xchg(&cmap->cpu_map[key_cpu], rcpu);
|
old_rcpu = unrcu_pointer(xchg(&cmap->cpu_map[key_cpu], RCU_INITIALIZER(rcpu)));
|
||||||
if (old_rcpu) {
|
if (old_rcpu) {
|
||||||
call_rcu(&old_rcpu->rcu, __cpu_map_entry_free);
|
call_rcu(&old_rcpu->rcu, __cpu_map_entry_free);
|
||||||
INIT_WORK(&old_rcpu->kthread_stop_wq, cpu_map_kthread_stop);
|
INIT_WORK(&old_rcpu->kthread_stop_wq, cpu_map_kthread_stop);
|
||||||
@ -551,7 +551,7 @@ static void cpu_map_free(struct bpf_map *map)
|
|||||||
for (i = 0; i < cmap->map.max_entries; i++) {
|
for (i = 0; i < cmap->map.max_entries; i++) {
|
||||||
struct bpf_cpu_map_entry *rcpu;
|
struct bpf_cpu_map_entry *rcpu;
|
||||||
|
|
||||||
rcpu = READ_ONCE(cmap->cpu_map[i]);
|
rcpu = rcu_dereference_raw(cmap->cpu_map[i]);
|
||||||
if (!rcpu)
|
if (!rcpu)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
@ -562,6 +562,10 @@ static void cpu_map_free(struct bpf_map *map)
|
|||||||
kfree(cmap);
|
kfree(cmap);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Elements are kept alive by RCU; either by rcu_read_lock() (from syscall) or
|
||||||
|
* by local_bh_disable() (from XDP calls inside NAPI). The
|
||||||
|
* rcu_read_lock_bh_held() below makes lockdep accept both.
|
||||||
|
*/
|
||||||
static void *__cpu_map_lookup_elem(struct bpf_map *map, u32 key)
|
static void *__cpu_map_lookup_elem(struct bpf_map *map, u32 key)
|
||||||
{
|
{
|
||||||
struct bpf_cpu_map *cmap = container_of(map, struct bpf_cpu_map, map);
|
struct bpf_cpu_map *cmap = container_of(map, struct bpf_cpu_map, map);
|
||||||
@ -570,7 +574,8 @@ static void *__cpu_map_lookup_elem(struct bpf_map *map, u32 key)
|
|||||||
if (key >= map->max_entries)
|
if (key >= map->max_entries)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
rcpu = READ_ONCE(cmap->cpu_map[key]);
|
rcpu = rcu_dereference_check(cmap->cpu_map[key],
|
||||||
|
rcu_read_lock_bh_held());
|
||||||
return rcpu;
|
return rcpu;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -73,7 +73,7 @@ struct bpf_dtab_netdev {
|
|||||||
|
|
||||||
struct bpf_dtab {
|
struct bpf_dtab {
|
||||||
struct bpf_map map;
|
struct bpf_map map;
|
||||||
struct bpf_dtab_netdev **netdev_map; /* DEVMAP type only */
|
struct bpf_dtab_netdev __rcu **netdev_map; /* DEVMAP type only */
|
||||||
struct list_head list;
|
struct list_head list;
|
||||||
|
|
||||||
/* these are only used for DEVMAP_HASH type maps */
|
/* these are only used for DEVMAP_HASH type maps */
|
||||||
@ -226,7 +226,7 @@ static void dev_map_free(struct bpf_map *map)
|
|||||||
for (i = 0; i < dtab->map.max_entries; i++) {
|
for (i = 0; i < dtab->map.max_entries; i++) {
|
||||||
struct bpf_dtab_netdev *dev;
|
struct bpf_dtab_netdev *dev;
|
||||||
|
|
||||||
dev = dtab->netdev_map[i];
|
dev = rcu_dereference_raw(dtab->netdev_map[i]);
|
||||||
if (!dev)
|
if (!dev)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
@ -259,6 +259,10 @@ static int dev_map_get_next_key(struct bpf_map *map, void *key, void *next_key)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Elements are kept alive by RCU; either by rcu_read_lock() (from syscall) or
|
||||||
|
* by local_bh_disable() (from XDP calls inside NAPI). The
|
||||||
|
* rcu_read_lock_bh_held() below makes lockdep accept both.
|
||||||
|
*/
|
||||||
static void *__dev_map_hash_lookup_elem(struct bpf_map *map, u32 key)
|
static void *__dev_map_hash_lookup_elem(struct bpf_map *map, u32 key)
|
||||||
{
|
{
|
||||||
struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map);
|
struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map);
|
||||||
@ -410,15 +414,9 @@ static void bq_xmit_all(struct xdp_dev_bulk_queue *bq, u32 flags)
|
|||||||
trace_xdp_devmap_xmit(bq->dev_rx, dev, sent, cnt - sent, err);
|
trace_xdp_devmap_xmit(bq->dev_rx, dev, sent, cnt - sent, err);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* __dev_flush is called from xdp_do_flush() which _must_ be signaled
|
/* __dev_flush is called from xdp_do_flush() which _must_ be signalled from the
|
||||||
* from the driver before returning from its napi->poll() routine. The poll()
|
* driver before returning from its napi->poll() routine. See the comment above
|
||||||
* routine is called either from busy_poll context or net_rx_action signaled
|
* xdp_do_flush() in filter.c.
|
||||||
* from NET_RX_SOFTIRQ. Either way the poll routine must complete before the
|
|
||||||
* net device can be torn down. On devmap tear down we ensure the flush list
|
|
||||||
* is empty before completing to ensure all flush operations have completed.
|
|
||||||
* When drivers update the bpf program they may need to ensure any flush ops
|
|
||||||
* are also complete. Using synchronize_rcu or call_rcu will suffice for this
|
|
||||||
* because both wait for napi context to exit.
|
|
||||||
*/
|
*/
|
||||||
void __dev_flush(void)
|
void __dev_flush(void)
|
||||||
{
|
{
|
||||||
@ -433,9 +431,9 @@ void __dev_flush(void)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/* rcu_read_lock (from syscall and BPF contexts) ensures that if a delete and/or
|
/* Elements are kept alive by RCU; either by rcu_read_lock() (from syscall) or
|
||||||
* update happens in parallel here a dev_put won't happen until after reading
|
* by local_bh_disable() (from XDP calls inside NAPI). The
|
||||||
* the ifindex.
|
* rcu_read_lock_bh_held() below makes lockdep accept both.
|
||||||
*/
|
*/
|
||||||
static void *__dev_map_lookup_elem(struct bpf_map *map, u32 key)
|
static void *__dev_map_lookup_elem(struct bpf_map *map, u32 key)
|
||||||
{
|
{
|
||||||
@ -445,12 +443,14 @@ static void *__dev_map_lookup_elem(struct bpf_map *map, u32 key)
|
|||||||
if (key >= map->max_entries)
|
if (key >= map->max_entries)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
obj = READ_ONCE(dtab->netdev_map[key]);
|
obj = rcu_dereference_check(dtab->netdev_map[key],
|
||||||
|
rcu_read_lock_bh_held());
|
||||||
return obj;
|
return obj;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Runs under RCU-read-side, plus in softirq under NAPI protection.
|
/* Runs in NAPI, i.e., softirq under local_bh_disable(). Thus, safe percpu
|
||||||
* Thus, safe percpu variable access.
|
* variable access, and map elements stick around. See comment above
|
||||||
|
* xdp_do_flush() in filter.c.
|
||||||
*/
|
*/
|
||||||
static void bq_enqueue(struct net_device *dev, struct xdp_frame *xdpf,
|
static void bq_enqueue(struct net_device *dev, struct xdp_frame *xdpf,
|
||||||
struct net_device *dev_rx, struct bpf_prog *xdp_prog)
|
struct net_device *dev_rx, struct bpf_prog *xdp_prog)
|
||||||
@ -735,14 +735,7 @@ static int dev_map_delete_elem(struct bpf_map *map, void *key)
|
|||||||
if (k >= map->max_entries)
|
if (k >= map->max_entries)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
/* Use call_rcu() here to ensure any rcu critical sections have
|
old_dev = unrcu_pointer(xchg(&dtab->netdev_map[k], NULL));
|
||||||
* completed as well as any flush operations because call_rcu
|
|
||||||
* will wait for preempt-disable region to complete, NAPI in this
|
|
||||||
* context. And additionally, the driver tear down ensures all
|
|
||||||
* soft irqs are complete before removing the net device in the
|
|
||||||
* case of dev_put equals zero.
|
|
||||||
*/
|
|
||||||
old_dev = xchg(&dtab->netdev_map[k], NULL);
|
|
||||||
if (old_dev)
|
if (old_dev)
|
||||||
call_rcu(&old_dev->rcu, __dev_map_entry_free);
|
call_rcu(&old_dev->rcu, __dev_map_entry_free);
|
||||||
return 0;
|
return 0;
|
||||||
@ -851,7 +844,7 @@ static int __dev_map_update_elem(struct net *net, struct bpf_map *map,
|
|||||||
* Remembering the driver side flush operation will happen before the
|
* Remembering the driver side flush operation will happen before the
|
||||||
* net device is removed.
|
* net device is removed.
|
||||||
*/
|
*/
|
||||||
old_dev = xchg(&dtab->netdev_map[i], dev);
|
old_dev = unrcu_pointer(xchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev)));
|
||||||
if (old_dev)
|
if (old_dev)
|
||||||
call_rcu(&old_dev->rcu, __dev_map_entry_free);
|
call_rcu(&old_dev->rcu, __dev_map_entry_free);
|
||||||
|
|
||||||
@ -1031,10 +1024,10 @@ static int dev_map_notification(struct notifier_block *notifier,
|
|||||||
for (i = 0; i < dtab->map.max_entries; i++) {
|
for (i = 0; i < dtab->map.max_entries; i++) {
|
||||||
struct bpf_dtab_netdev *dev, *odev;
|
struct bpf_dtab_netdev *dev, *odev;
|
||||||
|
|
||||||
dev = READ_ONCE(dtab->netdev_map[i]);
|
dev = rcu_dereference(dtab->netdev_map[i]);
|
||||||
if (!dev || netdev != dev->dev)
|
if (!dev || netdev != dev->dev)
|
||||||
continue;
|
continue;
|
||||||
odev = cmpxchg(&dtab->netdev_map[i], dev, NULL);
|
odev = unrcu_pointer(cmpxchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev), NULL));
|
||||||
if (dev == odev)
|
if (dev == odev)
|
||||||
call_rcu(&dev->rcu,
|
call_rcu(&dev->rcu,
|
||||||
__dev_map_entry_free);
|
__dev_map_entry_free);
|
||||||
|
@ -3897,6 +3897,34 @@ static const struct bpf_func_proto bpf_xdp_adjust_meta_proto = {
|
|||||||
.arg2_type = ARG_ANYTHING,
|
.arg2_type = ARG_ANYTHING,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
/* XDP_REDIRECT works by a three-step process, implemented in the functions
|
||||||
|
* below:
|
||||||
|
*
|
||||||
|
* 1. The bpf_redirect() and bpf_redirect_map() helpers will lookup the target
|
||||||
|
* of the redirect and store it (along with some other metadata) in a per-CPU
|
||||||
|
* struct bpf_redirect_info.
|
||||||
|
*
|
||||||
|
* 2. When the program returns the XDP_REDIRECT return code, the driver will
|
||||||
|
* call xdp_do_redirect() which will use the information in struct
|
||||||
|
* bpf_redirect_info to actually enqueue the frame into a map type-specific
|
||||||
|
* bulk queue structure.
|
||||||
|
*
|
||||||
|
* 3. Before exiting its NAPI poll loop, the driver will call xdp_do_flush(),
|
||||||
|
* which will flush all the different bulk queues, thus completing the
|
||||||
|
* redirect.
|
||||||
|
*
|
||||||
|
* Pointers to the map entries will be kept around for this whole sequence of
|
||||||
|
* steps, protected by RCU. However, there is no top-level rcu_read_lock() in
|
||||||
|
* the core code; instead, the RCU protection relies on everything happening
|
||||||
|
* inside a single NAPI poll sequence, which means it's between a pair of calls
|
||||||
|
* to local_bh_disable()/local_bh_enable().
|
||||||
|
*
|
||||||
|
* The map entries are marked as __rcu and the map code makes sure to
|
||||||
|
* dereference those pointers with rcu_dereference_check() in a way that works
|
||||||
|
* for both sections that to hold an rcu_read_lock() and sections that are
|
||||||
|
* called from NAPI without a separate rcu_read_lock(). The code below does not
|
||||||
|
* use RCU annotations, but relies on those in the map code.
|
||||||
|
*/
|
||||||
void xdp_do_flush(void)
|
void xdp_do_flush(void)
|
||||||
{
|
{
|
||||||
__dev_flush();
|
__dev_flush();
|
||||||
|
@ -749,7 +749,7 @@ static void xsk_unbind_dev(struct xdp_sock *xs)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static struct xsk_map *xsk_get_map_list_entry(struct xdp_sock *xs,
|
static struct xsk_map *xsk_get_map_list_entry(struct xdp_sock *xs,
|
||||||
struct xdp_sock ***map_entry)
|
struct xdp_sock __rcu ***map_entry)
|
||||||
{
|
{
|
||||||
struct xsk_map *map = NULL;
|
struct xsk_map *map = NULL;
|
||||||
struct xsk_map_node *node;
|
struct xsk_map_node *node;
|
||||||
@ -785,7 +785,7 @@ static void xsk_delete_from_maps(struct xdp_sock *xs)
|
|||||||
* might be updates to the map between
|
* might be updates to the map between
|
||||||
* xsk_get_map_list_entry() and xsk_map_try_sock_delete().
|
* xsk_get_map_list_entry() and xsk_map_try_sock_delete().
|
||||||
*/
|
*/
|
||||||
struct xdp_sock **map_entry = NULL;
|
struct xdp_sock __rcu **map_entry = NULL;
|
||||||
struct xsk_map *map;
|
struct xsk_map *map;
|
||||||
|
|
||||||
while ((map = xsk_get_map_list_entry(xs, &map_entry))) {
|
while ((map = xsk_get_map_list_entry(xs, &map_entry))) {
|
||||||
|
@ -31,7 +31,7 @@ struct xdp_mmap_offsets_v1 {
|
|||||||
struct xsk_map_node {
|
struct xsk_map_node {
|
||||||
struct list_head node;
|
struct list_head node;
|
||||||
struct xsk_map *map;
|
struct xsk_map *map;
|
||||||
struct xdp_sock **map_entry;
|
struct xdp_sock __rcu **map_entry;
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline struct xdp_sock *xdp_sk(struct sock *sk)
|
static inline struct xdp_sock *xdp_sk(struct sock *sk)
|
||||||
@ -40,7 +40,7 @@ static inline struct xdp_sock *xdp_sk(struct sock *sk)
|
|||||||
}
|
}
|
||||||
|
|
||||||
void xsk_map_try_sock_delete(struct xsk_map *map, struct xdp_sock *xs,
|
void xsk_map_try_sock_delete(struct xsk_map *map, struct xdp_sock *xs,
|
||||||
struct xdp_sock **map_entry);
|
struct xdp_sock __rcu **map_entry);
|
||||||
void xsk_clear_pool_at_qid(struct net_device *dev, u16 queue_id);
|
void xsk_clear_pool_at_qid(struct net_device *dev, u16 queue_id);
|
||||||
int xsk_reg_pool_at_qid(struct net_device *dev, struct xsk_buff_pool *pool,
|
int xsk_reg_pool_at_qid(struct net_device *dev, struct xsk_buff_pool *pool,
|
||||||
u16 queue_id);
|
u16 queue_id);
|
||||||
|
@ -12,7 +12,7 @@
|
|||||||
#include "xsk.h"
|
#include "xsk.h"
|
||||||
|
|
||||||
static struct xsk_map_node *xsk_map_node_alloc(struct xsk_map *map,
|
static struct xsk_map_node *xsk_map_node_alloc(struct xsk_map *map,
|
||||||
struct xdp_sock **map_entry)
|
struct xdp_sock __rcu **map_entry)
|
||||||
{
|
{
|
||||||
struct xsk_map_node *node;
|
struct xsk_map_node *node;
|
||||||
|
|
||||||
@ -42,7 +42,7 @@ static void xsk_map_sock_add(struct xdp_sock *xs, struct xsk_map_node *node)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void xsk_map_sock_delete(struct xdp_sock *xs,
|
static void xsk_map_sock_delete(struct xdp_sock *xs,
|
||||||
struct xdp_sock **map_entry)
|
struct xdp_sock __rcu **map_entry)
|
||||||
{
|
{
|
||||||
struct xsk_map_node *n, *tmp;
|
struct xsk_map_node *n, *tmp;
|
||||||
|
|
||||||
@ -124,6 +124,10 @@ static int xsk_map_gen_lookup(struct bpf_map *map, struct bpf_insn *insn_buf)
|
|||||||
return insn - insn_buf;
|
return insn - insn_buf;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Elements are kept alive by RCU; either by rcu_read_lock() (from syscall) or
|
||||||
|
* by local_bh_disable() (from XDP calls inside NAPI). The
|
||||||
|
* rcu_read_lock_bh_held() below makes lockdep accept both.
|
||||||
|
*/
|
||||||
static void *__xsk_map_lookup_elem(struct bpf_map *map, u32 key)
|
static void *__xsk_map_lookup_elem(struct bpf_map *map, u32 key)
|
||||||
{
|
{
|
||||||
struct xsk_map *m = container_of(map, struct xsk_map, map);
|
struct xsk_map *m = container_of(map, struct xsk_map, map);
|
||||||
@ -131,12 +135,11 @@ static void *__xsk_map_lookup_elem(struct bpf_map *map, u32 key)
|
|||||||
if (key >= map->max_entries)
|
if (key >= map->max_entries)
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
||||||
return READ_ONCE(m->xsk_map[key]);
|
return rcu_dereference_check(m->xsk_map[key], rcu_read_lock_bh_held());
|
||||||
}
|
}
|
||||||
|
|
||||||
static void *xsk_map_lookup_elem(struct bpf_map *map, void *key)
|
static void *xsk_map_lookup_elem(struct bpf_map *map, void *key)
|
||||||
{
|
{
|
||||||
WARN_ON_ONCE(!rcu_read_lock_held());
|
|
||||||
return __xsk_map_lookup_elem(map, *(u32 *)key);
|
return __xsk_map_lookup_elem(map, *(u32 *)key);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -149,7 +152,8 @@ static int xsk_map_update_elem(struct bpf_map *map, void *key, void *value,
|
|||||||
u64 map_flags)
|
u64 map_flags)
|
||||||
{
|
{
|
||||||
struct xsk_map *m = container_of(map, struct xsk_map, map);
|
struct xsk_map *m = container_of(map, struct xsk_map, map);
|
||||||
struct xdp_sock *xs, *old_xs, **map_entry;
|
struct xdp_sock __rcu **map_entry;
|
||||||
|
struct xdp_sock *xs, *old_xs;
|
||||||
u32 i = *(u32 *)key, fd = *(u32 *)value;
|
u32 i = *(u32 *)key, fd = *(u32 *)value;
|
||||||
struct xsk_map_node *node;
|
struct xsk_map_node *node;
|
||||||
struct socket *sock;
|
struct socket *sock;
|
||||||
@ -179,7 +183,7 @@ static int xsk_map_update_elem(struct bpf_map *map, void *key, void *value,
|
|||||||
}
|
}
|
||||||
|
|
||||||
spin_lock_bh(&m->lock);
|
spin_lock_bh(&m->lock);
|
||||||
old_xs = READ_ONCE(*map_entry);
|
old_xs = rcu_dereference_protected(*map_entry, lockdep_is_held(&m->lock));
|
||||||
if (old_xs == xs) {
|
if (old_xs == xs) {
|
||||||
err = 0;
|
err = 0;
|
||||||
goto out;
|
goto out;
|
||||||
@ -191,7 +195,7 @@ static int xsk_map_update_elem(struct bpf_map *map, void *key, void *value,
|
|||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
xsk_map_sock_add(xs, node);
|
xsk_map_sock_add(xs, node);
|
||||||
WRITE_ONCE(*map_entry, xs);
|
rcu_assign_pointer(*map_entry, xs);
|
||||||
if (old_xs)
|
if (old_xs)
|
||||||
xsk_map_sock_delete(old_xs, map_entry);
|
xsk_map_sock_delete(old_xs, map_entry);
|
||||||
spin_unlock_bh(&m->lock);
|
spin_unlock_bh(&m->lock);
|
||||||
@ -208,7 +212,8 @@ static int xsk_map_update_elem(struct bpf_map *map, void *key, void *value,
|
|||||||
static int xsk_map_delete_elem(struct bpf_map *map, void *key)
|
static int xsk_map_delete_elem(struct bpf_map *map, void *key)
|
||||||
{
|
{
|
||||||
struct xsk_map *m = container_of(map, struct xsk_map, map);
|
struct xsk_map *m = container_of(map, struct xsk_map, map);
|
||||||
struct xdp_sock *old_xs, **map_entry;
|
struct xdp_sock __rcu **map_entry;
|
||||||
|
struct xdp_sock *old_xs;
|
||||||
int k = *(u32 *)key;
|
int k = *(u32 *)key;
|
||||||
|
|
||||||
if (k >= map->max_entries)
|
if (k >= map->max_entries)
|
||||||
@ -216,7 +221,7 @@ static int xsk_map_delete_elem(struct bpf_map *map, void *key)
|
|||||||
|
|
||||||
spin_lock_bh(&m->lock);
|
spin_lock_bh(&m->lock);
|
||||||
map_entry = &m->xsk_map[k];
|
map_entry = &m->xsk_map[k];
|
||||||
old_xs = xchg(map_entry, NULL);
|
old_xs = unrcu_pointer(xchg(map_entry, NULL));
|
||||||
if (old_xs)
|
if (old_xs)
|
||||||
xsk_map_sock_delete(old_xs, map_entry);
|
xsk_map_sock_delete(old_xs, map_entry);
|
||||||
spin_unlock_bh(&m->lock);
|
spin_unlock_bh(&m->lock);
|
||||||
@ -231,11 +236,11 @@ static int xsk_map_redirect(struct bpf_map *map, u32 ifindex, u64 flags)
|
|||||||
}
|
}
|
||||||
|
|
||||||
void xsk_map_try_sock_delete(struct xsk_map *map, struct xdp_sock *xs,
|
void xsk_map_try_sock_delete(struct xsk_map *map, struct xdp_sock *xs,
|
||||||
struct xdp_sock **map_entry)
|
struct xdp_sock __rcu **map_entry)
|
||||||
{
|
{
|
||||||
spin_lock_bh(&map->lock);
|
spin_lock_bh(&map->lock);
|
||||||
if (READ_ONCE(*map_entry) == xs) {
|
if (rcu_access_pointer(*map_entry) == xs) {
|
||||||
WRITE_ONCE(*map_entry, NULL);
|
rcu_assign_pointer(*map_entry, NULL);
|
||||||
xsk_map_sock_delete(xs, map_entry);
|
xsk_map_sock_delete(xs, map_entry);
|
||||||
}
|
}
|
||||||
spin_unlock_bh(&map->lock);
|
spin_unlock_bh(&map->lock);
|
||||||
|
Loading…
Reference in New Issue
Block a user