mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2024-12-28 16:53:49 +00:00
0f92140468
Implement a memory provider that allocates dmabuf devmem in the form of net_iov. The provider receives a reference to the struct netdev_dmabuf_binding via the pool->mp_priv pointer. The driver needs to set this pointer for the provider in the net_iov. The provider obtains a reference on the netdev_dmabuf_binding which guarantees the binding and the underlying mapping remains alive until the provider is destroyed. Usage of PP_FLAG_DMA_MAP is required for this memory provide such that the page_pool can provide the driver with the dma-addrs of the devmem. Support for PP_FLAG_DMA_SYNC_DEV is omitted for simplicity & p.order != 0. Signed-off-by: Willem de Bruijn <willemb@google.com> Signed-off-by: Kaiyuan Zhang <kaiyuanz@google.com> Signed-off-by: Mina Almasry <almasrymina@google.com> Reviewed-by: Pavel Begunkov <asml.silence@gmail.com> Reviewed-by: Jakub Kicinski <kuba@kernel.org> Link: https://patch.msgid.link/20240910171458.219195-7-almasrymina@google.com Signed-off-by: Jakub Kicinski <kuba@kernel.org>
59 lines
1.5 KiB
C
59 lines
1.5 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
|
|
#ifndef __PAGE_POOL_PRIV_H
|
|
#define __PAGE_POOL_PRIV_H
|
|
|
|
#include <net/page_pool/helpers.h>
|
|
|
|
#include "netmem_priv.h"
|
|
|
|
s32 page_pool_inflight(const struct page_pool *pool, bool strict);
|
|
|
|
int page_pool_list(struct page_pool *pool);
|
|
void page_pool_detached(struct page_pool *pool);
|
|
void page_pool_unlist(struct page_pool *pool);
|
|
|
|
static inline bool
|
|
page_pool_set_dma_addr_netmem(netmem_ref netmem, dma_addr_t addr)
|
|
{
|
|
if (PAGE_POOL_32BIT_ARCH_WITH_64BIT_DMA) {
|
|
netmem_set_dma_addr(netmem, addr >> PAGE_SHIFT);
|
|
|
|
/* We assume page alignment to shave off bottom bits,
|
|
* if this "compression" doesn't work we need to drop.
|
|
*/
|
|
return addr != (dma_addr_t)netmem_get_dma_addr(netmem)
|
|
<< PAGE_SHIFT;
|
|
}
|
|
|
|
netmem_set_dma_addr(netmem, addr);
|
|
return false;
|
|
}
|
|
|
|
static inline bool page_pool_set_dma_addr(struct page *page, dma_addr_t addr)
|
|
{
|
|
return page_pool_set_dma_addr_netmem(page_to_netmem(page), addr);
|
|
}
|
|
|
|
#if defined(CONFIG_PAGE_POOL)
|
|
void page_pool_set_pp_info(struct page_pool *pool, netmem_ref netmem);
|
|
void page_pool_clear_pp_info(netmem_ref netmem);
|
|
int page_pool_check_memory_provider(struct net_device *dev,
|
|
struct netdev_rx_queue *rxq);
|
|
#else
|
|
static inline void page_pool_set_pp_info(struct page_pool *pool,
|
|
netmem_ref netmem)
|
|
{
|
|
}
|
|
static inline void page_pool_clear_pp_info(netmem_ref netmem)
|
|
{
|
|
}
|
|
static inline int page_pool_check_memory_provider(struct net_device *dev,
|
|
struct netdev_rx_queue *rxq)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#endif
|