Hello, On 21.10.2019 15:28, Madalin-cristian Bucur wrote: > The DPAA Ethernet driver is using the FMan MAC as the device for DMA > mapping. This is not actually correct, as the real DMA device is the > FMan port (the FMan Rx port for reception and the FMan Tx port for > transmission). Changing the device used for DMA mapping to the Fman > Rx and Tx port devices. > > Signed-off-by: Madalin Bucur <madalin.bu...@nxp.com> > Signed-off-by: Laurentiu Tudor <laurentiu.tu...@nxp.com> > --- > drivers/net/ethernet/freescale/dpaa/dpaa_eth.c | 105 > +++++++++++++------------ > drivers/net/ethernet/freescale/dpaa/dpaa_eth.h | 8 +- > 2 files changed, 62 insertions(+), 51 deletions(-) > > diff --git a/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c > b/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c > index 8d5686d88d30..639cafaa59b8 100644 > --- a/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c > +++ b/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c > @@ -1335,15 +1335,15 @@ static void dpaa_fd_release(const struct net_device > *net_dev, > vaddr = phys_to_virt(qm_fd_addr(fd)); > sgt = vaddr + qm_fd_get_offset(fd); > > - dma_unmap_single(dpaa_bp->dev, qm_fd_addr(fd), dpaa_bp->size, > - DMA_FROM_DEVICE); > + dma_unmap_single(dpaa_bp->priv->rx_dma_dev, qm_fd_addr(fd), > + dpaa_bp->size, DMA_FROM_DEVICE); > > dpaa_release_sgt_members(sgt); > > - addr = dma_map_single(dpaa_bp->dev, vaddr, dpaa_bp->size, > - DMA_FROM_DEVICE); > - if (dma_mapping_error(dpaa_bp->dev, addr)) { > - dev_err(dpaa_bp->dev, "DMA mapping failed"); > + addr = dma_map_single(dpaa_bp->priv->rx_dma_dev, vaddr, > + dpaa_bp->size, DMA_FROM_DEVICE); > + if (dma_mapping_error(dpaa_bp->priv->rx_dma_dev, addr)) { > + netdev_err(net_dev, "DMA mapping failed"); > return; > } > bm_buffer_set64(&bmb, addr); > @@ -1488,7 +1488,7 @@ static int dpaa_enable_tx_csum(struct dpaa_priv *priv, > > static int dpaa_bp_add_8_bufs(const struct dpaa_bp *dpaa_bp) > { > - struct device *dev = dpaa_bp->dev; > + struct net_device *net_dev = dpaa_bp->priv->net_dev; > struct bm_buffer bmb[8]; > dma_addr_t addr; > void *new_buf; > @@ -1497,16 +1497,18 @@ static int dpaa_bp_add_8_bufs(const struct dpaa_bp > *dpaa_bp) > for (i = 0; i < 8; i++) { > new_buf = netdev_alloc_frag(dpaa_bp->raw_size); > if (unlikely(!new_buf)) { > - dev_err(dev, "netdev_alloc_frag() failed, size %zu\n", > - dpaa_bp->raw_size); > + netdev_err(net_dev, > + "netdev_alloc_frag() failed, size %zu\n", > + dpaa_bp->raw_size); > goto release_previous_buffs; > } > new_buf = PTR_ALIGN(new_buf, SMP_CACHE_BYTES); > > - addr = dma_map_single(dev, new_buf, > + addr = dma_map_single(dpaa_bp->priv->rx_dma_dev, new_buf, > dpaa_bp->size, DMA_FROM_DEVICE); > - if (unlikely(dma_mapping_error(dev, addr))) { > - dev_err(dpaa_bp->dev, "DMA map failed"); > + if (unlikely(dma_mapping_error(dpaa_bp->priv->rx_dma_dev, > + addr))) { > + netdev_err(net_dev, "DMA map failed"); > goto release_previous_buffs; > } > > @@ -1634,7 +1636,7 @@ static struct sk_buff *dpaa_cleanup_tx_fd(const struct > dpaa_priv *priv, > > if (unlikely(qm_fd_get_format(fd) == qm_fd_sg)) { > nr_frags = skb_shinfo(skb)->nr_frags; > - dma_unmap_single(dev, addr, > + dma_unmap_single(priv->tx_dma_dev, addr, > qm_fd_get_offset(fd) + DPAA_SGT_SIZE, > dma_dir); > > @@ -1644,21 +1646,21 @@ static struct sk_buff *dpaa_cleanup_tx_fd(const > struct dpaa_priv *priv, > sgt = phys_to_virt(addr + qm_fd_get_offset(fd)); > > /* sgt[0] is from lowmem, was dma_map_single()-ed */ > - dma_unmap_single(dev, qm_sg_addr(&sgt[0]), > + dma_unmap_single(priv->tx_dma_dev, qm_sg_addr(&sgt[0]), > qm_sg_entry_get_len(&sgt[0]), dma_dir); > > /* remaining pages were mapped with skb_frag_dma_map() */ > for (i = 1; i <= nr_frags; i++) { > WARN_ON(qm_sg_entry_is_ext(&sgt[i])); > > - dma_unmap_page(dev, qm_sg_addr(&sgt[i]), > + dma_unmap_page(priv->tx_dma_dev, qm_sg_addr(&sgt[i]), > qm_sg_entry_get_len(&sgt[i]), dma_dir); > } > > /* Free the page frag that we allocated on Tx */ > skb_free_frag(phys_to_virt(addr)); > } else { > - dma_unmap_single(dev, addr, > + dma_unmap_single(priv->tx_dma_dev, addr, > skb_tail_pointer(skb) - (u8 *)skbh, dma_dir); > } > > @@ -1762,8 +1764,8 @@ static struct sk_buff *sg_fd_to_skb(const struct > dpaa_priv *priv, > goto free_buffers; > > count_ptr = this_cpu_ptr(dpaa_bp->percpu_count); > - dma_unmap_single(dpaa_bp->dev, sg_addr, dpaa_bp->size, > - DMA_FROM_DEVICE); > + dma_unmap_single(dpaa_bp->priv->rx_dma_dev, sg_addr, > + dpaa_bp->size, DMA_FROM_DEVICE); > if (!skb) { > sz = dpaa_bp->size + > SKB_DATA_ALIGN(sizeof(struct skb_shared_info)); > @@ -1853,7 +1855,6 @@ static int skb_to_contig_fd(struct dpaa_priv *priv, > int *offset) > { > struct net_device *net_dev = priv->net_dev; > - struct device *dev = net_dev->dev.parent; > enum dma_data_direction dma_dir; > unsigned char *buffer_start; > struct sk_buff **skbh; > @@ -1889,9 +1890,9 @@ static int skb_to_contig_fd(struct dpaa_priv *priv, > fd->cmd |= cpu_to_be32(FM_FD_CMD_FCO); > > /* Map the entire buffer size that may be seen by FMan, but no more */ > - addr = dma_map_single(dev, skbh, > + addr = dma_map_single(priv->tx_dma_dev, skbh, > skb_tail_pointer(skb) - buffer_start, dma_dir); > - if (unlikely(dma_mapping_error(dev, addr))) { > + if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) { > if (net_ratelimit()) > netif_err(priv, tx_err, net_dev, "dma_map_single() > failed\n"); > return -EINVAL; > @@ -1907,7 +1908,6 @@ static int skb_to_sg_fd(struct dpaa_priv *priv, > const enum dma_data_direction dma_dir = DMA_TO_DEVICE; > const int nr_frags = skb_shinfo(skb)->nr_frags; > struct net_device *net_dev = priv->net_dev; > - struct device *dev = net_dev->dev.parent; > struct qm_sg_entry *sgt; > struct sk_buff **skbh; > int i, j, err, sz; > @@ -1946,10 +1946,10 @@ static int skb_to_sg_fd(struct dpaa_priv *priv, > qm_sg_entry_set_len(&sgt[0], frag_len); > sgt[0].bpid = FSL_DPAA_BPID_INV; > sgt[0].offset = 0; > - addr = dma_map_single(dev, skb->data, > + addr = dma_map_single(priv->tx_dma_dev, skb->data, > skb_headlen(skb), dma_dir); > - if (unlikely(dma_mapping_error(dev, addr))) { > - dev_err(dev, "DMA mapping failed"); > + if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) { > + netdev_err(priv->net_dev, "DMA mapping failed"); > err = -EINVAL; > goto sg0_map_failed; > } > @@ -1960,10 +1960,10 @@ static int skb_to_sg_fd(struct dpaa_priv *priv, > frag = &skb_shinfo(skb)->frags[i]; > frag_len = skb_frag_size(frag); > WARN_ON(!skb_frag_page(frag)); > - addr = skb_frag_dma_map(dev, frag, 0, > + addr = skb_frag_dma_map(priv->tx_dma_dev, frag, 0, > frag_len, dma_dir); > - if (unlikely(dma_mapping_error(dev, addr))) { > - dev_err(dev, "DMA mapping failed"); > + if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) { > + netdev_err(priv->net_dev, "DMA mapping failed"); > err = -EINVAL; > goto sg_map_failed; > } > @@ -1986,10 +1986,10 @@ static int skb_to_sg_fd(struct dpaa_priv *priv, > skbh = (struct sk_buff **)buffer_start; > *skbh = skb; > > - addr = dma_map_single(dev, buffer_start, > + addr = dma_map_single(priv->tx_dma_dev, buffer_start, > priv->tx_headroom + DPAA_SGT_SIZE, dma_dir); > - if (unlikely(dma_mapping_error(dev, addr))) { > - dev_err(dev, "DMA mapping failed"); > + if (unlikely(dma_mapping_error(priv->tx_dma_dev, addr))) { > + netdev_err(priv->net_dev, "DMA mapping failed"); > err = -EINVAL; > goto sgt_map_failed; > } > @@ -2003,7 +2003,7 @@ static int skb_to_sg_fd(struct dpaa_priv *priv, > sgt_map_failed: > sg_map_failed: > for (j = 0; j < i; j++) > - dma_unmap_page(dev, qm_sg_addr(&sgt[j]), > + dma_unmap_page(priv->tx_dma_dev, qm_sg_addr(&sgt[j]), > qm_sg_entry_get_len(&sgt[j]), dma_dir); > sg0_map_failed: > csum_failed: > @@ -2304,7 +2304,8 @@ static enum qman_cb_dqrr_result rx_default_dqrr(struct > qman_portal *portal, > return qman_cb_dqrr_consume; > } > > - dma_unmap_single(dpaa_bp->dev, addr, dpaa_bp->size, DMA_FROM_DEVICE); > + dma_unmap_single(dpaa_bp->priv->rx_dma_dev, addr, dpaa_bp->size, > + DMA_FROM_DEVICE); > > /* prefetch the first 64 bytes of the frame or the SGT start */ > vaddr = phys_to_virt(addr); > @@ -2659,7 +2660,7 @@ static inline void dpaa_bp_free_pf(const struct dpaa_bp > *bp, > { > dma_addr_t addr = bm_buf_addr(bmb); > > - dma_unmap_single(bp->dev, addr, bp->size, DMA_FROM_DEVICE); > + dma_unmap_single(bp->priv->rx_dma_dev, addr, bp->size, DMA_FROM_DEVICE); > > skb_free_frag(phys_to_virt(addr)); > } > @@ -2769,25 +2770,27 @@ static int dpaa_eth_probe(struct platform_device > *pdev) > int err = 0, i, channel; > struct device *dev; > > + dev = &pdev->dev; > + > err = bman_is_probed(); > if (!err) > return -EPROBE_DEFER; > if (err < 0) { > - dev_err(&pdev->dev, "failing probe due to bman probe error\n"); > + dev_err(dev, "failing probe due to bman probe error\n");
These changes seem unrelated. > return -ENODEV; > } > err = qman_is_probed(); > if (!err) > return -EPROBE_DEFER; > if (err < 0) { > - dev_err(&pdev->dev, "failing probe due to qman probe error\n"); > + dev_err(dev, "failing probe due to qman probe error\n"); > return -ENODEV; > } > err = bman_portals_probed(); > if (!err) > return -EPROBE_DEFER; > if (err < 0) { > - dev_err(&pdev->dev, > + dev_err(dev, > "failing probe due to bman portals probe error\n"); > return -ENODEV; > } > @@ -2795,19 +2798,11 @@ static int dpaa_eth_probe(struct platform_device > *pdev) > if (!err) > return -EPROBE_DEFER; > if (err < 0) { > - dev_err(&pdev->dev, > + dev_err(dev, > "failing probe due to qman portals probe error\n"); > return -ENODEV; > } > > - /* device used for DMA mapping */ > - dev = pdev->dev.parent; > - err = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(40)); > - if (err) { > - dev_err(dev, "dma_coerce_mask_and_coherent() failed\n"); > - return err; > - } Why are we dropping this explicit setting of the dma mask? --- Best Regards, Laurentiu