On Sat, 29 Sep 2018 14:28:01 +0300 Ilias Apalodimas <ilias.apalodi...@linaro.org> wrote:
> +static void *netsec_alloc_rx_data(struct netsec_priv *priv, > + dma_addr_t *dma_handle, u16 *desc_len) > +{ > + size_t len = priv->ndev->mtu + ETH_HLEN + 2 * VLAN_HLEN + NET_SKB_PAD + > + NET_IP_ALIGN; > + dma_addr_t mapping; > + void *buf; > + > + len += SKB_DATA_ALIGN(sizeof(struct skb_shared_info)); > + len = SKB_DATA_ALIGN(len); > + > + buf = napi_alloc_frag(len); Using napi_alloc_frag here ^^^^ > + if (!buf) > + return NULL; > + > + mapping = dma_map_single(priv->dev, buf, len, DMA_FROM_DEVICE); > + if (unlikely(dma_mapping_error(priv->dev, mapping))) > + goto err_out; > + > + *dma_handle = mapping; > + *desc_len = len; > + > + return buf; > + > +err_out: > + skb_free_frag(buf); > + return NULL; > +} Hmmm, you are using napi_alloc_frag() in above code, which behind your-back allocates order-3 pages (32 Kbytes memory in 8 order-0 pages). This violates at-least two XDP principals: #1: You are NOT using order-0 page based allocations for XDP. Notice, I'm not saying 1-page per packet, as ixgbe + i40e violated this, and it is now "per-practical-code-example" acceptable to split up the order-0 page, and store two RX frames per order-0 page (4096 bytes). (To make this fit you have to reduce XDP_HEADROOM to 192 bytes, which killed the idea of placing the SKB in this area). #2: You have allocations on the XDP fast-path. The REAL secret behind the XDP performance is to avoid allocations on the fast-path. While I just told you to use the page-allocator and order-0 pages, this will actually kill performance. Thus, to make this fast, you need a driver local recycle scheme that avoids going through the page allocator, which makes XDP_DROP and XDP_TX extremely fast. For the XDP_REDIRECT action (which you seems to be interested in, as this is needed for AF_XDP), there is a xdp_return_frame() API that can make this fast. To avoid every driver inventing their own driver local page-recycle cache (which many does today), we/I have created the page pool API. See include/net/page_pool.h, and look at how mlx5 driver uses it in v4.18 links[1][2][3]. Do notice, that mlx5 ALSO have a driver recycle scheme on top, which Tariq is working on removing or generalizing. AND also that mlx5 does not use the DMA mapping feature that page_pool also provide yet. (Contact me if you want to use page_pool for handing DMA mapping, we might need to export __page_pool_clean_page and call it before XDP_PASS action). [1] https://github.com/torvalds/linux/blob/v4.18/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c#L226 [2] https://github.com/torvalds/linux/blob/v4.18/drivers/net/ethernet/mellanox/mlx5/core/en_rx.c#L255 [3] https://github.com/torvalds/linux/blob/v4.18/drivers/net/ethernet/mellanox/mlx5/core/en_main.c#L598-L618 -- Best regards, Jesper Dangaard Brouer MSc.CS, Principal Kernel Engineer at Red Hat LinkedIn: http://www.linkedin.com/in/brouer