// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB
|
/* Copyright (c) 2019 Mellanox Technologies. */
|
|
#include "rx.h"
|
#include "en/xdp.h"
|
#include <net/xdp_sock_drv.h>
|
|
/* RX data path */
|
|
static struct sk_buff *mlx5e_xsk_construct_skb(struct mlx5e_rq *rq, void *data,
|
u32 cqe_bcnt)
|
{
|
struct sk_buff *skb;
|
|
skb = napi_alloc_skb(rq->cq.napi, cqe_bcnt);
|
if (unlikely(!skb)) {
|
rq->stats->buff_alloc_err++;
|
return NULL;
|
}
|
|
skb_put_data(skb, data, cqe_bcnt);
|
|
return skb;
|
}
|
|
struct sk_buff *mlx5e_xsk_skb_from_cqe_mpwrq_linear(struct mlx5e_rq *rq,
|
struct mlx5e_mpw_info *wi,
|
u16 cqe_bcnt,
|
u32 head_offset,
|
u32 page_idx)
|
{
|
struct xdp_buff *xdp = wi->umr.dma_info[page_idx].xsk;
|
u32 cqe_bcnt32 = cqe_bcnt;
|
|
/* Check packet size. Note LRO doesn't use linear SKB */
|
if (unlikely(cqe_bcnt > rq->hw_mtu)) {
|
rq->stats->oversize_pkts_sw_drop++;
|
return NULL;
|
}
|
|
/* head_offset is not used in this function, because xdp->data and the
|
* DMA address point directly to the necessary place. Furthermore, in
|
* the current implementation, UMR pages are mapped to XSK frames, so
|
* head_offset should always be 0.
|
*/
|
WARN_ON_ONCE(head_offset);
|
|
xdp->data_end = xdp->data + cqe_bcnt32;
|
xdp_set_data_meta_invalid(xdp);
|
xsk_buff_dma_sync_for_cpu(xdp, rq->xsk_pool);
|
net_prefetch(xdp->data);
|
|
/* Possible flows:
|
* - XDP_REDIRECT to XSKMAP:
|
* The page is owned by the userspace from now.
|
* - XDP_TX and other XDP_REDIRECTs:
|
* The page was returned by ZCA and recycled.
|
* - XDP_DROP:
|
* Recycle the page.
|
* - XDP_PASS:
|
* Allocate an SKB, copy the data and recycle the page.
|
*
|
* Pages to be recycled go to the Reuse Ring on MPWQE deallocation. Its
|
* size is the same as the Driver RX Ring's size, and pages for WQEs are
|
* allocated first from the Reuse Ring, so it has enough space.
|
*/
|
|
if (likely(mlx5e_xdp_handle(rq, NULL, &cqe_bcnt32, xdp))) {
|
if (likely(__test_and_clear_bit(MLX5E_RQ_FLAG_XDP_XMIT, rq->flags)))
|
__set_bit(page_idx, wi->xdp_xmit_bitmap); /* non-atomic */
|
return NULL; /* page/packet was consumed by XDP */
|
}
|
|
/* XDP_PASS: copy the data from the UMEM to a new SKB and reuse the
|
* frame. On SKB allocation failure, NULL is returned.
|
*/
|
return mlx5e_xsk_construct_skb(rq, xdp->data, cqe_bcnt32);
|
}
|
|
struct sk_buff *mlx5e_xsk_skb_from_cqe_linear(struct mlx5e_rq *rq,
|
struct mlx5_cqe64 *cqe,
|
struct mlx5e_wqe_frag_info *wi,
|
u32 cqe_bcnt)
|
{
|
struct xdp_buff *xdp = wi->di->xsk;
|
|
/* wi->offset is not used in this function, because xdp->data and the
|
* DMA address point directly to the necessary place. Furthermore, the
|
* XSK allocator allocates frames per packet, instead of pages, so
|
* wi->offset should always be 0.
|
*/
|
WARN_ON_ONCE(wi->offset);
|
|
xdp->data_end = xdp->data + cqe_bcnt;
|
xdp_set_data_meta_invalid(xdp);
|
xsk_buff_dma_sync_for_cpu(xdp, rq->xsk_pool);
|
net_prefetch(xdp->data);
|
|
if (unlikely(get_cqe_opcode(cqe) != MLX5_CQE_RESP_SEND)) {
|
rq->stats->wqe_err++;
|
return NULL;
|
}
|
|
if (likely(mlx5e_xdp_handle(rq, NULL, &cqe_bcnt, xdp)))
|
return NULL; /* page/packet was consumed by XDP */
|
|
/* XDP_PASS: copy the data from the UMEM to a new SKB. The frame reuse
|
* will be handled by mlx5e_put_rx_frag.
|
* On SKB allocation failure, NULL is returned.
|
*/
|
return mlx5e_xsk_construct_skb(rq, xdp->data, cqe_bcnt);
|
}
|