On Tue, 24 Jun 2025 18:14:16 +0000
Morten Brørup <m...@smartsharesystems.com> wrote:

> Added fast mbuf release, re-using the existing mbuf pool pointer
> in the queue structure.
> 
> Signed-off-by: Morten Brørup <m...@smartsharesystems.com>

Makes sense.

> ---
>  drivers/net/null/rte_eth_null.c | 30 +++++++++++++++++++++++++++---
>  1 file changed, 27 insertions(+), 3 deletions(-)
> 
> diff --git a/drivers/net/null/rte_eth_null.c b/drivers/net/null/rte_eth_null.c
> index 8a9b74a03b..12c0d8d1ff 100644
> --- a/drivers/net/null/rte_eth_null.c
> +++ b/drivers/net/null/rte_eth_null.c
> @@ -34,6 +34,17 @@ struct pmd_internals;
>  struct null_queue {
>       struct pmd_internals *internals;
>  
> +     /**
> +      * For RX queue:
> +      *  Mempool to allocate mbufs from.
> +      *
> +      * For TX queue:
> +      *  Mempool to free mbufs to, if fast release of mbufs is enabled.
> +      *  UINTPTR_MAX if the mempool for fast release of mbufs has not yet 
> been detected.
> +      *  NULL if fast release of mbufs is not enabled.
> +      *
> +      *  @see RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE
> +      */
>       struct rte_mempool *mb_pool;

Do all drivers to it this way?
Is it documented in ethdev?

>       void *dummy_packet;
>  
> @@ -151,7 +162,16 @@ eth_null_tx(void *q, struct rte_mbuf **bufs, uint16_t 
> nb_bufs)
>       for (i = 0; i < nb_bufs; i++)
>               bytes += rte_pktmbuf_pkt_len(bufs[i]);
>  
> -     rte_pktmbuf_free_bulk(bufs, nb_bufs);
> +     if (h->mb_pool != NULL) { /* RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE */
> +             if (unlikely(h->mb_pool == (void *)UINTPTR_MAX)) {
> +                     if (unlikely(nb_bufs == 0))
> +                             return 0; /* Do not dereference uninitialized 
> bufs[0]. */
> +                     h->mb_pool = bufs[0]->pool;
> +             }
> +             rte_mbuf_raw_free_bulk(h->mb_pool, bufs, nb_bufs);
> +     } else {
> +             rte_pktmbuf_free_bulk(bufs, nb_bufs);
> +     }
>       rte_atomic_fetch_add_explicit(&h->tx_pkts, nb_bufs, 
> rte_memory_order_relaxed);
>       rte_atomic_fetch_add_explicit(&h->tx_bytes, bytes, 
> rte_memory_order_relaxed);
>  
> @@ -259,7 +279,7 @@ static int
>  eth_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id,
>               uint16_t nb_tx_desc __rte_unused,
>               unsigned int socket_id __rte_unused,
> -             const struct rte_eth_txconf *tx_conf __rte_unused)
> +             const struct rte_eth_txconf *tx_conf)
>  {
>       struct rte_mbuf *dummy_packet;
>       struct pmd_internals *internals;
> @@ -284,6 +304,9 @@ eth_tx_queue_setup(struct rte_eth_dev *dev, uint16_t 
> tx_queue_id,
>  
>       internals->tx_null_queues[tx_queue_id].internals = internals;
>       internals->tx_null_queues[tx_queue_id].dummy_packet = dummy_packet;
> +     internals->tx_null_queues[tx_queue_id].mb_pool =
> +                     tx_conf->offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE ?
> +                     (void *)UINTPTR_MAX : NULL;
>  
>       return 0;
>  }
> @@ -309,7 +332,8 @@ eth_dev_info(struct rte_eth_dev *dev,
>       dev_info->max_rx_queues = RTE_DIM(internals->rx_null_queues);
>       dev_info->max_tx_queues = RTE_DIM(internals->tx_null_queues);
>       dev_info->min_rx_bufsize = 0;
> -     dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS | 
> RTE_ETH_TX_OFFLOAD_MT_LOCKFREE;
> +     dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS |
> +                     RTE_ETH_TX_OFFLOAD_MT_LOCKFREE | 
> RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE;
>  
>       dev_info->reta_size = internals->reta_size;
>       dev_info->flow_type_rss_offloads = internals->flow_type_rss_offloads;

Reply via email to