added RTE_PMD_DLB2_ prefix to dlb2 token pop mode enmus to avoid name comflict. These enums are passed to public API rte_pmd_dlb2_set_token_pop_mode().
Fixes: c667583d82f4 ("event/dlb2: add token pop API") Cc: sta...@dpdk.org Signed-off-by: Pravin Pathak <pravin.pat...@intel.com> --- drivers/event/dlb2/dlb2.c | 28 +++++++++++++++------------- drivers/event/dlb2/dlb2_priv.h | 2 +- drivers/event/dlb2/dlb2_selftest.c | 6 +++--- drivers/event/dlb2/rte_pmd_dlb2.c | 4 ++-- drivers/event/dlb2/rte_pmd_dlb2.h | 23 ++++++++++++----------- 5 files changed, 33 insertions(+), 30 deletions(-) diff --git a/drivers/event/dlb2/dlb2.c b/drivers/event/dlb2/dlb2.c index fd8cc70f3c..084875f1c8 100644 --- a/drivers/event/dlb2/dlb2.c +++ b/drivers/event/dlb2/dlb2.c @@ -1819,7 +1819,7 @@ dlb2_hw_create_ldb_port(struct dlb2_eventdev *dlb2, /* The default enqueue functions do not include delayed-pop support for * performance reasons. */ - if (qm_port->token_pop_mode == DELAYED_POP) { + if (qm_port->token_pop_mode == RTE_PMD_DLB2_DELAYED_POP) { dlb2->event_dev->enqueue_burst = dlb2_event_enqueue_burst_delayed; dlb2->event_dev->enqueue_new_burst = @@ -2021,7 +2021,7 @@ dlb2_hw_create_dir_port(struct dlb2_eventdev *dlb2, qm_port->dequeue_depth = dequeue_depth; /* Directed ports are auto-pop, by default. */ - qm_port->token_pop_mode = AUTO_POP; + qm_port->token_pop_mode = RTE_PMD_DLB2_AUTO_POP; qm_port->owed_tokens = 0; qm_port->issued_releases = 0; @@ -3359,7 +3359,7 @@ __dlb2_event_enqueue_burst_reorder(void *event_port, } } - if (use_delayed && qm_port->token_pop_mode == DELAYED_POP && + if (use_delayed && qm_port->token_pop_mode == RTE_PMD_DLB2_DELAYED_POP && (events[i].op == RTE_EVENT_OP_FORWARD || events[i].op == RTE_EVENT_OP_RELEASE) && qm_port->issued_releases >= thresh - 1) { @@ -3468,7 +3468,7 @@ __dlb2_event_enqueue_burst(void *event_port, int ret; if (use_delayed && - qm_port->token_pop_mode == DELAYED_POP && + qm_port->token_pop_mode == RTE_PMD_DLB2_DELAYED_POP && (ev->op == RTE_EVENT_OP_FORWARD || ev->op == RTE_EVENT_OP_RELEASE) && qm_port->issued_releases >= thresh - 1) { @@ -3620,7 +3620,7 @@ dlb2_event_release(struct dlb2_eventdev *dlb2, for (; j < DLB2_NUM_QES_PER_CACHE_LINE && (i + j) < n; j++) { int16_t thresh = qm_port->token_pop_thresh; - if (qm_port->token_pop_mode == DELAYED_POP && + if (qm_port->token_pop_mode == RTE_PMD_DLB2_DELAYED_POP && qm_port->issued_releases >= thresh - 1) { /* Insert the token pop QE */ dlb2_construct_token_pop_qe(qm_port, j); @@ -4365,7 +4365,7 @@ dlb2_hw_dequeue_sparse(struct dlb2_eventdev *dlb2, qm_port->owed_tokens += num; if (num) { - if (qm_port->token_pop_mode == AUTO_POP) + if (qm_port->token_pop_mode == RTE_PMD_DLB2_AUTO_POP) dlb2_consume_qe_immediate(qm_port, num); ev_port->outstanding_releases += num; @@ -4495,7 +4495,7 @@ dlb2_hw_dequeue(struct dlb2_eventdev *dlb2, qm_port->owed_tokens += num; if (num) { - if (qm_port->token_pop_mode == AUTO_POP) + if (qm_port->token_pop_mode == RTE_PMD_DLB2_AUTO_POP) dlb2_consume_qe_immediate(qm_port, num); ev_port->outstanding_releases += num; @@ -4540,7 +4540,7 @@ dlb2_event_dequeue_burst(void *event_port, struct rte_event *ev, uint16_t num, order->enq_reorder[i].u64[1] = release_u64; __dlb2_event_enqueue_burst_reorder(event_port, NULL, 0, - qm_port->token_pop_mode == DELAYED_POP); + qm_port->token_pop_mode == RTE_PMD_DLB2_DELAYED_POP); } else { dlb2_event_release(dlb2, ev_port->id, out_rels); } @@ -4548,7 +4548,7 @@ dlb2_event_dequeue_burst(void *event_port, struct rte_event *ev, uint16_t num, DLB2_INC_STAT(ev_port->stats.tx_implicit_rel, out_rels); } - if (qm_port->token_pop_mode == DEFERRED_POP && qm_port->owed_tokens) + if (qm_port->token_pop_mode == RTE_PMD_DLB2_DEFERRED_POP && qm_port->owed_tokens) dlb2_consume_qe_immediate(qm_port, qm_port->owed_tokens); cnt = dlb2_hw_dequeue(dlb2, ev_port, ev, num, wait); @@ -4597,7 +4597,8 @@ dlb2_event_dequeue_burst_sparse(void *event_port, struct rte_event *ev, if (num_releases == RTE_DIM(release_burst)) { __dlb2_event_enqueue_burst_reorder(event_port, release_burst, RTE_DIM(release_burst), - qm_port->token_pop_mode == DELAYED_POP); + qm_port->token_pop_mode == + RTE_PMD_DLB2_DELAYED_POP); num_releases = 0; } } @@ -4605,7 +4606,8 @@ dlb2_event_dequeue_burst_sparse(void *event_port, struct rte_event *ev, if (num_releases) __dlb2_event_enqueue_burst_reorder(event_port, release_burst - , num_releases, qm_port->token_pop_mode == DELAYED_POP); + , num_releases, + qm_port->token_pop_mode == RTE_PMD_DLB2_DELAYED_POP); } else { dlb2_event_release(dlb2, ev_port->id, out_rels); } @@ -4614,7 +4616,7 @@ dlb2_event_dequeue_burst_sparse(void *event_port, struct rte_event *ev, DLB2_INC_STAT(ev_port->stats.tx_implicit_rel, out_rels); } - if (qm_port->token_pop_mode == DEFERRED_POP && qm_port->owed_tokens) + if (qm_port->token_pop_mode == RTE_PMD_DLB2_DEFERRED_POP && qm_port->owed_tokens) dlb2_consume_qe_immediate(qm_port, qm_port->owed_tokens); cnt = dlb2_hw_dequeue_sparse(dlb2, ev_port, ev, num, wait); @@ -5142,7 +5144,7 @@ dlb2_primary_eventdev_probe(struct rte_eventdev *dev, /* Initialize each port's token pop mode */ for (i = 0; i < DLB2_MAX_NUM_PORTS(dlb2->version); i++) - dlb2->ev_ports[i].qm_port.token_pop_mode = AUTO_POP; + dlb2->ev_ports[i].qm_port.token_pop_mode = RTE_PMD_DLB2_AUTO_POP; rte_spinlock_init(&dlb2->qm_instance.resource_lock); diff --git a/drivers/event/dlb2/dlb2_priv.h b/drivers/event/dlb2/dlb2_priv.h index 30d1d5b9ae..7a5cbcca1e 100644 --- a/drivers/event/dlb2/dlb2_priv.h +++ b/drivers/event/dlb2/dlb2_priv.h @@ -349,7 +349,7 @@ struct dlb2_port { bool gen_bit; uint16_t dir_credits; uint32_t dequeue_depth; - enum dlb2_token_pop_mode token_pop_mode; + enum rte_pmd_dlb2_token_pop_mode token_pop_mode; union dlb2_port_config cfg; RTE_ATOMIC(uint32_t) *credit_pool[DLB2_NUM_QUEUE_TYPES]; union { diff --git a/drivers/event/dlb2/dlb2_selftest.c b/drivers/event/dlb2/dlb2_selftest.c index 62aa11d981..87d98700c9 100644 --- a/drivers/event/dlb2/dlb2_selftest.c +++ b/drivers/event/dlb2/dlb2_selftest.c @@ -1105,13 +1105,13 @@ test_deferred_sched(void) return -1; } - ret = rte_pmd_dlb2_set_token_pop_mode(evdev, 0, DEFERRED_POP); + ret = rte_pmd_dlb2_set_token_pop_mode(evdev, 0, RTE_PMD_DLB2_DEFERRED_POP); if (ret < 0) { printf("%d: Error setting deferred scheduling\n", __LINE__); goto err; } - ret = rte_pmd_dlb2_set_token_pop_mode(evdev, 1, DEFERRED_POP); + ret = rte_pmd_dlb2_set_token_pop_mode(evdev, 1, RTE_PMD_DLB2_DEFERRED_POP); if (ret < 0) { printf("%d: Error setting deferred scheduling\n", __LINE__); goto err; @@ -1257,7 +1257,7 @@ test_delayed_pop(void) return -1; } - ret = rte_pmd_dlb2_set_token_pop_mode(evdev, 0, DELAYED_POP); + ret = rte_pmd_dlb2_set_token_pop_mode(evdev, 0, RTE_PMD_DLB2_DELAYED_POP); if (ret < 0) { printf("%d: Error setting deferred scheduling\n", __LINE__); goto err; diff --git a/drivers/event/dlb2/rte_pmd_dlb2.c b/drivers/event/dlb2/rte_pmd_dlb2.c index b75010027d..80186dd07d 100644 --- a/drivers/event/dlb2/rte_pmd_dlb2.c +++ b/drivers/event/dlb2/rte_pmd_dlb2.c @@ -14,7 +14,7 @@ RTE_EXPORT_EXPERIMENTAL_SYMBOL(rte_pmd_dlb2_set_token_pop_mode, 20.11) int rte_pmd_dlb2_set_token_pop_mode(uint8_t dev_id, uint8_t port_id, - enum dlb2_token_pop_mode mode) + enum rte_pmd_dlb2_token_pop_mode mode) { struct dlb2_eventdev *dlb2; struct rte_eventdev *dev; @@ -24,7 +24,7 @@ rte_pmd_dlb2_set_token_pop_mode(uint8_t dev_id, dlb2 = dlb2_pmd_priv(dev); - if (mode >= NUM_TOKEN_POP_MODES) + if (mode >= RTE_PMD_DLB2_NUM_TOKEN_POP_MODES) return -EINVAL; /* The event device must be configured, but not yet started */ diff --git a/drivers/event/dlb2/rte_pmd_dlb2.h b/drivers/event/dlb2/rte_pmd_dlb2.h index f58ef2168d..33e741261d 100644 --- a/drivers/event/dlb2/rte_pmd_dlb2.h +++ b/drivers/event/dlb2/rte_pmd_dlb2.h @@ -49,18 +49,18 @@ extern "C" { * * Selects the token pop mode for a DLB2 port. */ -enum dlb2_token_pop_mode { +enum rte_pmd_dlb2_token_pop_mode { /* Pop the CQ tokens immediately after dequeuing. */ - AUTO_POP, + RTE_PMD_DLB2_AUTO_POP, /* Pop CQ tokens after (dequeue_depth - 1) events are released. * Supported on load-balanced ports only. */ - DELAYED_POP, + RTE_PMD_DLB2_DELAYED_POP, /* Pop the CQ tokens during next dequeue operation. */ - DEFERRED_POP, + RTE_PMD_DLB2_DEFERRED_POP, /* NUM_TOKEN_POP_MODES must be last */ - NUM_TOKEN_POP_MODES + RTE_PMD_DLB2_NUM_TOKEN_POP_MODES }; /** @@ -68,8 +68,9 @@ enum dlb2_token_pop_mode { * @b EXPERIMENTAL: this API may change, or be removed, without prior notice * * Configure the token pop mode for a DLB2 port. By default, all ports use - * AUTO_POP. This function must be called before calling rte_event_port_setup() - * for the port, but after calling rte_event_dev_configure(). + * RTE_PMD_DLB2_AUTO_POP. This function must be called before calling + * rte_event_port_setup() for the port, but after calling + * rte_event_dev_configure(). * * @param dev_id * The identifier of the event device. @@ -80,16 +81,16 @@ enum dlb2_token_pop_mode { * * @return * - 0: Success - * - EINVAL: Invalid dev_id, port_id, or mode - * - EINVAL: The DLB2 is not configured, is already running, or the port is - * already setup + * - EINVAL: Invalid parameter dev_id, port_id, or mode + * - EINVAL: The DLB2 device is not configured or is already running, + * or the port is already setup */ __rte_experimental int rte_pmd_dlb2_set_token_pop_mode(uint8_t dev_id, uint8_t port_id, - enum dlb2_token_pop_mode mode); + enum rte_pmd_dlb2_token_pop_mode mode); /** Set inflight threshold for flow migration */ #define DLB2_SET_PORT_FLOW_MIGRATION_THRESHOLD RTE_BIT64(0) -- 2.39.1