Currently, there's a bunch of vector PMD-related stuff that's being compiled based on what architecture the code is being compiled for.
Simplify it by removing necessary #ifdef's and have it so that vector PMD compilation is controlled in one place - using a compile flag from the build file. Signed-off-by: Anatoly Burakov <anatoly.bura...@intel.com> --- Notes: v3 -> v4: - Add this commit drivers/net/intel/ixgbe/ixgbe_rxtx.c | 27 +++++++++++++++------------ drivers/net/intel/ixgbe/ixgbe_rxtx.h | 4 ---- drivers/net/intel/ixgbe/meson.build | 2 ++ 3 files changed, 17 insertions(+), 16 deletions(-) diff --git a/drivers/net/intel/ixgbe/ixgbe_rxtx.c b/drivers/net/intel/ixgbe/ixgbe_rxtx.c index f75821029d..f5fd50584a 100644 --- a/drivers/net/intel/ixgbe/ixgbe_rxtx.c +++ b/drivers/net/intel/ixgbe/ixgbe_rxtx.c @@ -2678,9 +2678,7 @@ ixgbe_set_tx_function(struct rte_eth_dev *dev, struct ci_tx_queue *txq) (rte_eal_process_type() != RTE_PROC_PRIMARY || ixgbe_txq_vec_setup(txq) == 0)) { PMD_INIT_LOG(DEBUG, "Vector tx enabled."); -#if defined(RTE_ARCH_X86) || defined(RTE_ARCH_ARM) dev->recycle_tx_mbufs_reuse = ixgbe_recycle_tx_mbufs_reuse_vec; -#endif dev->tx_pkt_burst = ixgbe_xmit_pkts_vec; } else dev->tx_pkt_burst = ixgbe_xmit_pkts_simple; @@ -5049,10 +5047,8 @@ ixgbe_set_rx_function(struct rte_eth_dev *dev) PMD_INIT_LOG(DEBUG, "Using Vector Scattered Rx " "callback (port=%d).", dev->data->port_id); -#if defined(RTE_ARCH_X86) || defined(RTE_ARCH_ARM) dev->recycle_rx_descriptors_refill = ixgbe_recycle_rx_descriptors_refill_vec; -#endif dev->rx_pkt_burst = ixgbe_recv_scattered_pkts_vec; } else if (adapter->rx_bulk_alloc_allowed) { PMD_INIT_LOG(DEBUG, "Using a Scattered with bulk " @@ -5081,9 +5077,7 @@ ixgbe_set_rx_function(struct rte_eth_dev *dev) "burst size no less than %d (port=%d).", RTE_IXGBE_DESCS_PER_LOOP, dev->data->port_id); -#if defined(RTE_ARCH_X86) || defined(RTE_ARCH_ARM) dev->recycle_rx_descriptors_refill = ixgbe_recycle_rx_descriptors_refill_vec; -#endif dev->rx_pkt_burst = ixgbe_recv_pkts_vec; } else if (adapter->rx_bulk_alloc_allowed) { PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions are " @@ -5871,10 +5865,8 @@ ixgbe_recycle_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, recycle_rxq_info->receive_tail = &rxq->rx_tail; if (adapter->rx_vec_allowed) { -#if defined(RTE_ARCH_X86) || defined(RTE_ARCH_ARM) recycle_rxq_info->refill_requirement = RTE_IXGBE_RXQ_REARM_THRESH; recycle_rxq_info->refill_head = &rxq->rxrearm_start; -#endif } else { recycle_rxq_info->refill_requirement = rxq->rx_free_thresh; recycle_rxq_info->refill_head = &rxq->rx_free_trigger; @@ -6239,11 +6231,9 @@ ixgbe_config_rss_filter(struct rte_eth_dev *dev, return 0; } -/* Stubs needed for linkage when RTE_ARCH_PPC_64, RTE_ARCH_RISCV or - * RTE_ARCH_LOONGARCH is set. +/* Stubs needed for linkage when vectorized PMD isn't supported. */ -#if defined(RTE_ARCH_PPC_64) || defined(RTE_ARCH_RISCV) || \ - defined(RTE_ARCH_LOONGARCH) +#ifndef IXGBE_VPMD_SUPPORTED int ixgbe_rx_vec_dev_conf_condition_check(struct rte_eth_dev __rte_unused *dev) { @@ -6268,6 +6258,12 @@ ixgbe_recv_scattered_pkts_vec( return 0; } +void +ixgbe_recycle_rx_descriptors_refill_vec(void __rte_unused * rx_queue, + uint16_t __rte_unused nb_mbufs) +{ +} + int ixgbe_rxq_vec_setup(struct ixgbe_rx_queue __rte_unused *rxq) { @@ -6282,6 +6278,13 @@ ixgbe_xmit_fixed_burst_vec(void __rte_unused *tx_queue, return 0; } +uint16_t +ixgbe_recycle_tx_mbufs_reuse_vec(void __rte_unused * tx_queue, + struct rte_eth_recycle_rxq_info __rte_unused * recycle_rxq_info) +{ + return 0; +} + int ixgbe_txq_vec_setup(struct ci_tx_queue *txq __rte_unused) { diff --git a/drivers/net/intel/ixgbe/ixgbe_rxtx.h b/drivers/net/intel/ixgbe/ixgbe_rxtx.h index c86714804f..bcd5db87e8 100644 --- a/drivers/net/intel/ixgbe/ixgbe_rxtx.h +++ b/drivers/net/intel/ixgbe/ixgbe_rxtx.h @@ -35,10 +35,8 @@ #define RTE_IXGBE_DESCS_PER_LOOP 4 -#if defined(RTE_ARCH_X86) || defined(RTE_ARCH_ARM) #define RTE_IXGBE_RXQ_REARM_THRESH 32 #define RTE_IXGBE_MAX_RX_BURST RTE_IXGBE_RXQ_REARM_THRESH -#endif #define RX_RING_SZ ((IXGBE_MAX_RING_DESC + RTE_PMD_IXGBE_RX_MAX_BURST) * \ sizeof(union ixgbe_adv_rx_desc)) @@ -102,10 +100,8 @@ struct ixgbe_rx_queue { uint8_t using_ipsec; /**< indicates that IPsec RX feature is in use */ #endif -#if defined(RTE_ARCH_X86) || defined(RTE_ARCH_ARM) uint16_t rxrearm_nb; /**< number of remaining to be re-armed */ uint16_t rxrearm_start; /**< the idx we start the re-arming from */ -#endif uint16_t rx_free_thresh; /**< max free RX desc to hold. */ uint16_t queue_id; /**< RX queue index. */ uint16_t reg_idx; /**< RX queue register index. */ diff --git a/drivers/net/intel/ixgbe/meson.build b/drivers/net/intel/ixgbe/meson.build index d1122bb9cd..8adb1567d1 100644 --- a/drivers/net/intel/ixgbe/meson.build +++ b/drivers/net/intel/ixgbe/meson.build @@ -26,9 +26,11 @@ deps += ['hash', 'security'] if arch_subdir == 'x86' sources += files('ixgbe_rxtx_vec_sse.c') sources += files('ixgbe_recycle_mbufs_vec_common.c') + cflags += ['-DIXGBE_VPMD_SUPPORTED'] elif arch_subdir == 'arm' sources += files('ixgbe_rxtx_vec_neon.c') sources += files('ixgbe_recycle_mbufs_vec_common.c') + cflags += ['-DIXGBE_VPMD_SUPPORTED'] endif headers = files('rte_pmd_ixgbe.h') -- 2.47.1