On 22.10.2019 12:10, Madalin-cristian Bucur wrote:
>> -----Original Message-----
>> From: Laurentiu Tudor
>> Sent: Tuesday, October 22, 2019 11:50 AM
>> To: Madalin-cristian Bucur <madalin.bu...@nxp.com>; da...@davemloft.net;
>> netdev@vger.kernel.org
>> Cc: Roy Pledge <roy.ple...@nxp.com>
>> Subject: Re: [PATCH net-next 5/6] dpaa_eth: change DMA device
>>
>> Hello,
>>
>> On 21.10.2019 15:28, Madalin-cristian Bucur wrote:
>>> The DPAA Ethernet driver is using the FMan MAC as the device for DMA
>>> mapping. This is not actually correct, as the real DMA device is the
>>> FMan port (the FMan Rx port for reception and the FMan Tx port for
>>> transmission). Changing the device used for DMA mapping to the Fman
>>> Rx and Tx port devices.
>>>
>>> Signed-off-by: Madalin Bucur <madalin.bu...@nxp.com>
>>> Signed-off-by: Laurentiu Tudor <laurentiu.tu...@nxp.com>
>>> ---
>>>    drivers/net/ethernet/freescale/dpaa/dpaa_eth.c | 105 +++++++++++++----
>> --------
>>>    drivers/net/ethernet/freescale/dpaa/dpaa_eth.h |   8 +-
>>>    2 files changed, 62 insertions(+), 51 deletions(-)
>>>
>>> diff --git a/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
>> b/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
>>> index 8d5686d88d30..639cafaa59b8 100644
>>> --- a/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
>>> +++ b/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
>>> @@ -1335,15 +1335,15 @@ static void dpaa_fd_release(const struct
>> net_device *net_dev,
>>>             vaddr = phys_to_virt(qm_fd_addr(fd));
>>>             sgt = vaddr + qm_fd_get_offset(fd);
>>>
>>> -           dma_unmap_single(dpaa_bp->dev, qm_fd_addr(fd), dpaa_bp->size,
>>> -                            DMA_FROM_DEVICE);
>>> +           dma_unmap_single(dpaa_bp->priv->rx_dma_dev, qm_fd_addr(fd),
>>> +                            dpaa_bp->size, DMA_FROM_DEVICE);
>>>
>>>             dpaa_release_sgt_members(sgt);
>>>
>>> -           addr = dma_map_single(dpaa_bp->dev, vaddr, dpaa_bp->size,
>>> -                                 DMA_FROM_DEVICE);
>>> -           if (dma_mapping_error(dpaa_bp->dev, addr)) {
>>> -                   dev_err(dpaa_bp->dev, "DMA mapping failed");
>>> +           addr = dma_map_single(dpaa_bp->priv->rx_dma_dev, vaddr,
>>> +                                 dpaa_bp->size, DMA_FROM_DEVICE);
>>> +           if (dma_mapping_error(dpaa_bp->priv->rx_dma_dev, addr)) {
>>> +                   netdev_err(net_dev, "DMA mapping failed");
>>>                     return;
>>>             }
>>>             bm_buffer_set64(&bmb, addr);
>>> @@ -1488,7 +1488,7 @@ static int dpaa_enable_tx_csum(struct dpaa_priv
>> *priv,
>>>
>>>    static int dpaa_bp_add_8_bufs(const struct dpaa_bp *dpaa_bp)
>>>    {
>>> -   struct device *dev = dpaa_bp->dev;
>>> +   struct net_device *net_dev = dpaa_bp->priv->net_dev;
>>>     struct bm_buffer bmb[8];
>>>     dma_addr_t addr;
>>>     void *new_buf;
>>> @@ -1497,16 +1497,18 @@ static int dpaa_bp_add_8_bufs(const struct
>> dpaa_bp *dpaa_bp)
>>>     for (i = 0; i < 8; i++) {
>>>             new_buf = netdev_alloc_frag(dpaa_bp->raw_size);
>>>             if (unlikely(!new_buf)) {
>>> -                   dev_err(dev, "netdev_alloc_frag() failed, size %zu\n",
>>> -                           dpaa_bp->raw_size);
>>> +                   netdev_err(net_dev,
>>> +                              "netdev_alloc_frag() failed, size %zu\n",
>>> +                              dpaa_bp->raw_size);
>>>                     goto release_previous_buffs;
>>>             }
>>>             new_buf = PTR_ALIGN(new_buf, SMP_CACHE_BYTES);
>>>
>>> -           addr = dma_map_single(dev, new_buf,
>>> +           addr = dma_map_single(dpaa_bp->priv->rx_dma_dev, new_buf,
>>>                                   dpaa_bp->size, DMA_FROM_DEVICE);
>>> -           if (unlikely(dma_mapping_error(dev, addr))) {
>>> -                   dev_err(dpaa_bp->dev, "DMA map failed");
>>> +           if (unlikely(dma_mapping_error(dpaa_bp->priv->rx_dma_dev,
>>> +                                          addr))) {
>>> +                   netdev_err(net_dev, "DMA map failed");
>>>                     goto release_previous_buffs;
>>>             }
>>>
>>> @@ -1634,7 +1636,7 @@ static struct sk_buff *dpaa_cleanup_tx_fd(const
>> struct dpaa_priv *priv,
>>>
>>>     if (unlikely(qm_fd_get_format(fd) == qm_fd_sg)) {
>>>             nr_frags = skb_shinfo(skb)->nr_frags;
>>> -           dma_unmap_single(dev, addr,
>>> +           dma_unmap_single(priv->tx_dma_dev, addr,
>>>                              qm_fd_get_offset(fd) + DPAA_SGT_SIZE,
>>>                              dma_dir);
>>>
>>> @@ -1644,21 +1646,21 @@ static struct sk_buff *dpaa_cleanup_tx_fd(const
>> struct dpaa_priv *priv,
>>>             sgt = phys_to_virt(addr + qm_fd_get_offset(fd));
>>>
>>>             /* sgt[0] is from lowmem, was dma_map_single()-ed */
>>> -           dma_unmap_single(dev, qm_sg_addr(&sgt[0]),
>>> +           dma_unmap_single(priv->tx_dma_dev, qm_sg_addr(&sgt[0]),
>>>                              qm_sg_entry_get_len(&sgt[0]), dma_dir);
>>>
>>>             /* remaining pages were mapped with skb_frag_dma_map() */
>>>             for (i = 1; i <= nr_frags; i++) {
>>>                     WARN_ON(qm_sg_entry_is_ext(&sgt[i]));
>>>
>>> -                   dma_unmap_page(dev, qm_sg_addr(&sgt[i]),
>>> +                   dma_unmap_page(priv->tx_dma_dev, qm_sg_addr(&sgt[i]),
>>>                                    qm_sg_entry_get_len(&sgt[i]), dma_dir);
>>>             }
>>>
>>>             /* Free the page frag that we allocated on Tx */
>>>             skb_free_frag(phys_to_virt(addr));
>>>     } else {
>>> -           dma_unmap_single(dev, addr,
>>> +           dma_unmap_single(priv->tx_dma_dev, addr,
>>>                              skb_tail_pointer(skb) - (u8 *)skbh, dma_dir);
>>>     }
>>>
>>> @@ -1762,8 +1764,8 @@ static struct sk_buff *sg_fd_to_skb(const struct
>> dpaa_priv *priv,
>>>                     goto free_buffers;
>>>
>>>             count_ptr = this_cpu_ptr(dpaa_bp->percpu_count);
>>> -           dma_unmap_single(dpaa_bp->dev, sg_addr, dpaa_bp->size,
>>> -                            DMA_FROM_DEVICE);
>>> +           dma_unmap_single(dpaa_bp->priv->rx_dma_dev, sg_addr,
>>> +                            dpaa_bp->size, DMA_FROM_DEVICE);
>>>             if (!skb) {
>>>                     sz = dpaa_bp->size +
>>>                             SKB_DATA_ALIGN(sizeof(struct skb_shared_info));
>>> @@ -1853,7 +1855,6 @@ static int skb_to_contig_fd(struct dpaa_priv
>> *priv,
>>>                         int *offset)
>>>    {
>>>     struct net_device *net_dev = priv->net_dev;
>>> -   struct device *dev = net_dev->dev.parent;
>>>     enum dma_data_direction dma_dir;
>>>     unsigned char *buffer_start;
>>>     struct sk_buff **skbh;
>>> @@ -1889,9 +1890,9 @@ static int skb_to_contig_fd(struct dpaa_priv
>> *priv,
>>>     fd->cmd |= cpu_to_be32(FM_FD_CMD_FCO);
>>>
>>>     /* Map the entire buffer size that may be seen by FMan, but no more
>> */
>>> -   addr = dma_map_single(dev, skbh,
>>> +   addr = dma_map_single(priv->tx_dma_dev, skbh,
>>>                           skb_tail_pointer(skb) - buffer_start, dma_dir);
>>> -   if (unlikely(dma_mapping_error(dev, addr))) {
>>> +   if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) {
>>>             if (net_ratelimit())
>>>                     netif_err(priv, tx_err, net_dev, "dma_map_single()
>> failed\n");
>>>             return -EINVAL;
>>> @@ -1907,7 +1908,6 @@ static int skb_to_sg_fd(struct dpaa_priv *priv,
>>>     const enum dma_data_direction dma_dir = DMA_TO_DEVICE;
>>>     const int nr_frags = skb_shinfo(skb)->nr_frags;
>>>     struct net_device *net_dev = priv->net_dev;
>>> -   struct device *dev = net_dev->dev.parent;
>>>     struct qm_sg_entry *sgt;
>>>     struct sk_buff **skbh;
>>>     int i, j, err, sz;
>>> @@ -1946,10 +1946,10 @@ static int skb_to_sg_fd(struct dpaa_priv *priv,
>>>     qm_sg_entry_set_len(&sgt[0], frag_len);
>>>     sgt[0].bpid = FSL_DPAA_BPID_INV;
>>>     sgt[0].offset = 0;
>>> -   addr = dma_map_single(dev, skb->data,
>>> +   addr = dma_map_single(priv->tx_dma_dev, skb->data,
>>>                           skb_headlen(skb), dma_dir);
>>> -   if (unlikely(dma_mapping_error(dev, addr))) {
>>> -           dev_err(dev, "DMA mapping failed");
>>> +   if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) {
>>> +           netdev_err(priv->net_dev, "DMA mapping failed");
>>>             err = -EINVAL;
>>>             goto sg0_map_failed;
>>>     }
>>> @@ -1960,10 +1960,10 @@ static int skb_to_sg_fd(struct dpaa_priv *priv,
>>>             frag = &skb_shinfo(skb)->frags[i];
>>>             frag_len = skb_frag_size(frag);
>>>             WARN_ON(!skb_frag_page(frag));
>>> -           addr = skb_frag_dma_map(dev, frag, 0,
>>> +           addr = skb_frag_dma_map(priv->tx_dma_dev, frag, 0,
>>>                                     frag_len, dma_dir);
>>> -           if (unlikely(dma_mapping_error(dev, addr))) {
>>> -                   dev_err(dev, "DMA mapping failed");
>>> +           if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) {
>>> +                   netdev_err(priv->net_dev, "DMA mapping failed");
>>>                     err = -EINVAL;
>>>                     goto sg_map_failed;
>>>             }
>>> @@ -1986,10 +1986,10 @@ static int skb_to_sg_fd(struct dpaa_priv *priv,
>>>     skbh = (struct sk_buff **)buffer_start;
>>>     *skbh = skb;
>>>
>>> -   addr = dma_map_single(dev, buffer_start,
>>> +   addr = dma_map_single(priv->tx_dma_dev, buffer_start,
>>>                           priv->tx_headroom + DPAA_SGT_SIZE, dma_dir);
>>> -   if (unlikely(dma_mapping_error(dev, addr))) {
>>> -           dev_err(dev, "DMA mapping failed");
>>> +   if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) {
>>> +           netdev_err(priv->net_dev, "DMA mapping failed");
>>>             err = -EINVAL;
>>>             goto sgt_map_failed;
>>>     }
>>> @@ -2003,7 +2003,7 @@ static int skb_to_sg_fd(struct dpaa_priv *priv,
>>>    sgt_map_failed:
>>>    sg_map_failed:
>>>     for (j = 0; j < i; j++)
>>> -           dma_unmap_page(dev, qm_sg_addr(&sgt[j]),
>>> +           dma_unmap_page(priv->tx_dma_dev, qm_sg_addr(&sgt[j]),
>>>                            qm_sg_entry_get_len(&sgt[j]), dma_dir);
>>>    sg0_map_failed:
>>>    csum_failed:
>>> @@ -2304,7 +2304,8 @@ static enum qman_cb_dqrr_result
>> rx_default_dqrr(struct qman_portal *portal,
>>>             return qman_cb_dqrr_consume;
>>>     }
>>>
>>> -   dma_unmap_single(dpaa_bp->dev, addr, dpaa_bp->size,
>> DMA_FROM_DEVICE);
>>> +   dma_unmap_single(dpaa_bp->priv->rx_dma_dev, addr, dpaa_bp->size,
>>> +                    DMA_FROM_DEVICE);
>>>
>>>     /* prefetch the first 64 bytes of the frame or the SGT start */
>>>     vaddr = phys_to_virt(addr);
>>> @@ -2659,7 +2660,7 @@ static inline void dpaa_bp_free_pf(const struct
>> dpaa_bp *bp,
>>>    {
>>>     dma_addr_t addr = bm_buf_addr(bmb);
>>>
>>> -   dma_unmap_single(bp->dev, addr, bp->size, DMA_FROM_DEVICE);
>>> +   dma_unmap_single(bp->priv->rx_dma_dev, addr, bp->size,
>> DMA_FROM_DEVICE);
>>>
>>>     skb_free_frag(phys_to_virt(addr));
>>>    }
>>> @@ -2769,25 +2770,27 @@ static int dpaa_eth_probe(struct platform_device
>> *pdev)
>>>     int err = 0, i, channel;
>>>     struct device *dev;
>>>
>>> +   dev = &pdev->dev;
>>> +
>>>     err = bman_is_probed();
>>>     if (!err)
>>>             return -EPROBE_DEFER;
>>>     if (err < 0) {
>>> -           dev_err(&pdev->dev, "failing probe due to bman probe
>> error\n");
>>> +           dev_err(dev, "failing probe due to bman probe error\n");
>>
>> These changes seem unrelated.
> 
> The &pdev->dev to dev replacement is not related directly to the incorrect
> DMA mapping device but a device had to be used for the prints and propagating
> &pdev->dev did not look like a good idea. Initially I had a separate patch for
> this but it's superfluous to add code and remove it in another patch doing 
> almost
> nothing.

Fair enough.

>>>             return -ENODEV;
>>>     }
>>>     err = qman_is_probed();
>>>     if (!err)
>>>             return -EPROBE_DEFER;
>>>     if (err < 0) {
>>> -           dev_err(&pdev->dev, "failing probe due to qman probe
>> error\n");
>>> +           dev_err(dev, "failing probe due to qman probe error\n");
>>>             return -ENODEV;
>>>     }
>>>     err = bman_portals_probed();
>>>     if (!err)
>>>             return -EPROBE_DEFER;
>>>     if (err < 0) {
>>> -           dev_err(&pdev->dev,
>>> +           dev_err(dev,
>>>                     "failing probe due to bman portals probe error\n");
>>>             return -ENODEV;
>>>     }
>>> @@ -2795,19 +2798,11 @@ static int dpaa_eth_probe(struct platform_device
>> *pdev)
>>>     if (!err)
>>>             return -EPROBE_DEFER;
>>>     if (err < 0) {
>>> -           dev_err(&pdev->dev,
>>> +           dev_err(dev,
>>>                     "failing probe due to qman portals probe error\n");
>>>             return -ENODEV;
>>>     }
>>>
>>> -   /* device used for DMA mapping */
>>> -   dev = pdev->dev.parent;
>>> -   err = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(40));
>>> -   if (err) {
>>> -           dev_err(dev, "dma_coerce_mask_and_coherent() failed\n");
>>> -           return err;
>>> -   }
>>
>> Why are we dropping this explicit setting of the dma mask?
>>
>> ---
>> Best Regards, Laurentiu
> 
> Hi Laurentiu, you are probably reviewing these changes with your initial
> patch in mind that was using (incorrectly) the same (Rx) port for DMA
> mapping of both receive and transmit traffic. Please take a second look at
> the changes in this patch.

My bad, didn't notice you're setting it below. Sorry for the noise.

---
Best Regards, Laurentiu

Reply via email to