From: Mark Starovoytov <mstarovoi...@marvell.com>

This patch adds u64_stats_update_* usage to protect access to 64-bit stats.

Signed-off-by: Mark Starovoytov <mstarovoi...@marvell.com>
Signed-off-by: Igor Russkikh <irussk...@marvell.com>
---
 .../net/ethernet/aquantia/atlantic/aq_ring.c  | 28 +++++++-
 .../net/ethernet/aquantia/atlantic/aq_ring.h  |  1 +
 .../net/ethernet/aquantia/atlantic/aq_vec.c   | 68 +++++++++++++------
 3 files changed, 77 insertions(+), 20 deletions(-)

diff --git a/drivers/net/ethernet/aquantia/atlantic/aq_ring.c 
b/drivers/net/ethernet/aquantia/atlantic/aq_ring.c
index 20b606aa7974..b073c340c005 100644
--- a/drivers/net/ethernet/aquantia/atlantic/aq_ring.c
+++ b/drivers/net/ethernet/aquantia/atlantic/aq_ring.c
@@ -70,26 +70,35 @@ static int aq_get_rxpages(struct aq_ring_s *self, struct 
aq_ring_buff_s *rxbuf,
                        rxbuf->rxdata.pg_off += AQ_CFG_RX_FRAME_MAX;
                        if (rxbuf->rxdata.pg_off + AQ_CFG_RX_FRAME_MAX <=
                                (PAGE_SIZE << order)) {
+                               u64_stats_update_begin(&self->stats.syncp);
                                self->stats.rx.pg_flips++;
+                               u64_stats_update_end(&self->stats.syncp);
                        } else {
                                /* Buffer exhausted. We have other users and
                                 * should release this page and realloc
                                 */
                                aq_free_rxpage(&rxbuf->rxdata,
                                               aq_nic_get_dev(self->aq_nic));
+                               u64_stats_update_begin(&self->stats.syncp);
                                self->stats.rx.pg_losts++;
+                               u64_stats_update_end(&self->stats.syncp);
                        }
                } else {
                        rxbuf->rxdata.pg_off = 0;
+                       u64_stats_update_begin(&self->stats.syncp);
                        self->stats.rx.pg_reuses++;
+                       u64_stats_update_end(&self->stats.syncp);
                }
        }
 
        if (!rxbuf->rxdata.page) {
                ret = aq_get_rxpage(&rxbuf->rxdata, order,
                                    aq_nic_get_dev(self->aq_nic));
-               if (ret)
+               if (ret) {
+                       u64_stats_update_begin(&self->stats.syncp);
                        self->stats.rx.alloc_fails++;
+                       u64_stats_update_end(&self->stats.syncp);
+               }
                return ret;
        }
 
@@ -213,6 +222,7 @@ int aq_ring_init(struct aq_ring_s *self)
        self->hw_head = 0;
        self->sw_head = 0;
        self->sw_tail = 0;
+       u64_stats_init(&self->stats.syncp);
 
        return 0;
 }
@@ -240,7 +250,9 @@ void aq_ring_queue_wake(struct aq_ring_s *ring)
                                                      ring->idx))) {
                netif_wake_subqueue(ndev,
                                    AQ_NIC_RING2QMAP(ring->aq_nic, ring->idx));
+               u64_stats_update_begin(&ring->stats.syncp);
                ring->stats.tx.queue_restarts++;
+               u64_stats_update_end(&ring->stats.syncp);
        }
 }
 
@@ -282,8 +294,10 @@ bool aq_ring_tx_clean(struct aq_ring_s *self)
                }
 
                if (unlikely(buff->is_eop)) {
+                       u64_stats_update_begin(&self->stats.syncp);
                        ++self->stats.tx.packets;
                        self->stats.tx.bytes += buff->skb->len;
+                       u64_stats_update_end(&self->stats.syncp);
 
                        dev_kfree_skb_any(buff->skb);
                }
@@ -303,7 +317,9 @@ static void aq_rx_checksum(struct aq_ring_s *self,
                return;
 
        if (unlikely(buff->is_cso_err)) {
+               u64_stats_update_begin(&self->stats.syncp);
                ++self->stats.rx.errors;
+               u64_stats_update_end(&self->stats.syncp);
                skb->ip_summed = CHECKSUM_NONE;
                return;
        }
@@ -373,13 +389,17 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
                                        buff_->is_cleaned = true;
                                } while (!buff_->is_eop);
 
+                               u64_stats_update_begin(&self->stats.syncp);
                                ++self->stats.rx.errors;
+                               u64_stats_update_end(&self->stats.syncp);
                                continue;
                        }
                }
 
                if (buff->is_error) {
+                       u64_stats_update_begin(&self->stats.syncp);
                        ++self->stats.rx.errors;
+                       u64_stats_update_end(&self->stats.syncp);
                        continue;
                }
 
@@ -394,7 +414,9 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
                        skb = build_skb(aq_buf_vaddr(&buff->rxdata),
                                        AQ_CFG_RX_FRAME_MAX);
                        if (unlikely(!skb)) {
+                               u64_stats_update_begin(&self->stats.syncp);
                                self->stats.rx.skb_alloc_fails++;
+                               u64_stats_update_end(&self->stats.syncp);
                                err = -ENOMEM;
                                goto err_exit;
                        }
@@ -408,7 +430,9 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
                } else {
                        skb = napi_alloc_skb(napi, AQ_CFG_RX_HDR_SIZE);
                        if (unlikely(!skb)) {
+                               u64_stats_update_begin(&self->stats.syncp);
                                self->stats.rx.skb_alloc_fails++;
+                               u64_stats_update_end(&self->stats.syncp);
                                err = -ENOMEM;
                                goto err_exit;
                        }
@@ -482,8 +506,10 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
                                                : AQ_NIC_RING2QMAP(self->aq_nic,
                                                                   self->idx));
 
+               u64_stats_update_begin(&self->stats.syncp);
                ++self->stats.rx.packets;
                self->stats.rx.bytes += skb->len;
+               u64_stats_update_end(&self->stats.syncp);
 
                napi_gro_receive(napi, skb);
        }
diff --git a/drivers/net/ethernet/aquantia/atlantic/aq_ring.h 
b/drivers/net/ethernet/aquantia/atlantic/aq_ring.h
index 0e40ccbaf19a..7ccdaeb957f7 100644
--- a/drivers/net/ethernet/aquantia/atlantic/aq_ring.h
+++ b/drivers/net/ethernet/aquantia/atlantic/aq_ring.h
@@ -110,6 +110,7 @@ struct aq_ring_stats_tx_s {
 };
 
 union aq_ring_stats_s {
+       struct u64_stats_sync syncp;
        struct aq_ring_stats_rx_s rx;
        struct aq_ring_stats_tx_s tx;
 };
diff --git a/drivers/net/ethernet/aquantia/atlantic/aq_vec.c 
b/drivers/net/ethernet/aquantia/atlantic/aq_vec.c
index eafcb98c274f..6539d2cc5459 100644
--- a/drivers/net/ethernet/aquantia/atlantic/aq_vec.c
+++ b/drivers/net/ethernet/aquantia/atlantic/aq_vec.c
@@ -45,7 +45,9 @@ static int aq_vec_poll(struct napi_struct *napi, int budget)
        } else {
                for (i = 0U, ring = self->ring[0];
                        self->tx_rings > i; ++i, ring = self->ring[i]) {
+                       u64_stats_update_begin(&ring[AQ_VEC_RX_ID].stats.syncp);
                        ring[AQ_VEC_RX_ID].stats.rx.polls++;
+                       u64_stats_update_end(&ring[AQ_VEC_RX_ID].stats.syncp);
                        if (self->aq_hw_ops->hw_ring_tx_head_update) {
                                err = self->aq_hw_ops->hw_ring_tx_head_update(
                                                        self->aq_hw,
@@ -357,29 +359,57 @@ void aq_vec_add_stats(struct aq_vec_s *self,
                      struct aq_ring_stats_tx_s *stats_tx)
 {
        struct aq_ring_s *ring = self->ring[tc];
+       u64 packets, bytes, errors;
+       unsigned int start;
 
        if (tc < self->rx_rings) {
-               struct aq_ring_stats_rx_s *rx = &ring[AQ_VEC_RX_ID].stats.rx;
-
-               stats_rx->packets += rx->packets;
-               stats_rx->bytes += rx->bytes;
-               stats_rx->errors += rx->errors;
-               stats_rx->jumbo_packets += rx->jumbo_packets;
-               stats_rx->lro_packets += rx->lro_packets;
-               stats_rx->alloc_fails += rx->alloc_fails;
-               stats_rx->skb_alloc_fails += rx->skb_alloc_fails;
-               stats_rx->polls += rx->polls;
-               stats_rx->pg_losts += rx->pg_losts;
-               stats_rx->pg_flips += rx->pg_flips;
-               stats_rx->pg_reuses += rx->pg_reuses;
+               u64 jumbo_packets, lro_packets, alloc_fails, skb_alloc_fails;
+               union aq_ring_stats_s *stats = &ring[AQ_VEC_RX_ID].stats;
+               u64 polls, pg_losts, pg_flips, pg_reuses;
+
+               do {
+                       start = u64_stats_fetch_begin_irq(&stats->syncp);
+                       packets = stats->rx.packets;
+                       bytes = stats->rx.bytes;
+                       errors = stats->rx.errors;
+                       jumbo_packets = stats->rx.jumbo_packets;
+                       lro_packets = stats->rx.lro_packets;
+                       alloc_fails = stats->rx.alloc_fails;
+                       skb_alloc_fails = stats->rx.skb_alloc_fails;
+                       polls = stats->rx.polls;
+                       pg_losts = stats->rx.pg_losts;
+                       pg_flips = stats->rx.pg_flips;
+                       pg_reuses = stats->rx.pg_reuses;
+               } while (u64_stats_fetch_retry_irq(&stats->syncp, start));
+
+               stats_rx->packets += packets;
+               stats_rx->bytes += bytes;
+               stats_rx->errors += errors;
+               stats_rx->jumbo_packets += jumbo_packets;
+               stats_rx->lro_packets += lro_packets;
+               stats_rx->alloc_fails += alloc_fails;
+               stats_rx->skb_alloc_fails += skb_alloc_fails;
+               stats_rx->polls += polls;
+               stats_rx->pg_losts += pg_losts;
+               stats_rx->pg_flips += pg_flips;
+               stats_rx->pg_reuses += pg_reuses;
        }
 
        if (tc < self->tx_rings) {
-               struct aq_ring_stats_tx_s *tx = &ring[AQ_VEC_TX_ID].stats.tx;
-
-               stats_tx->packets += tx->packets;
-               stats_tx->bytes += tx->bytes;
-               stats_tx->errors += tx->errors;
-               stats_tx->queue_restarts += tx->queue_restarts;
+               union aq_ring_stats_s *stats = &ring[AQ_VEC_TX_ID].stats;
+               u64 queue_restarts;
+
+               do {
+                       start = u64_stats_fetch_begin_irq(&stats->syncp);
+                       packets = stats->tx.packets;
+                       bytes = stats->tx.bytes;
+                       errors = stats->tx.errors;
+                       queue_restarts = stats->tx.queue_restarts;
+               } while (u64_stats_fetch_retry_irq(&stats->syncp, start));
+
+               stats_tx->packets += packets;
+               stats_tx->bytes += bytes;
+               stats_tx->errors += errors;
+               stats_tx->queue_restarts += queue_restarts;
        }
 }
-- 
2.25.1

Reply via email to