On Tue, 21 Mar 2023 16:45:21 -0400, "Michael S. Tsirkin" <[email protected]>
wrote:
> On Tue, Mar 21, 2023 at 05:34:59PM +0800, Xuan Zhuo wrote:
> > virtio core only supports virtual addresses, dma is completed in virtio
> > core.
> >
> > In some scenarios (such as the AF_XDP), the memory is allocated
> > and DMA mapping is completed in advance, so it is necessary for us to
> > support passing the DMA address to virtio core.
> >
> > Drives can use sg->dma_address to pass the mapped dma address to virtio
> > core. If one sg->dma_address is used then all sgs must use
> > sg->dma_address, otherwise all must be null when passing it to the APIs
> > of virtio.
> >
> > Signed-off-by: Xuan Zhuo <[email protected]>
> > ---
> > drivers/virtio/virtio_ring.c | 27 +++++++++++++++++++--------
> > 1 file changed, 19 insertions(+), 8 deletions(-)
> >
> > diff --git a/drivers/virtio/virtio_ring.c b/drivers/virtio/virtio_ring.c
> > index c8ed4aef9462..be2ff96964c3 100644
> > --- a/drivers/virtio/virtio_ring.c
> > +++ b/drivers/virtio/virtio_ring.c
> > @@ -70,6 +70,7 @@
> > struct vring_desc_state_split {
> > void *data; /* Data for callback. */
> > struct vring_desc *indir_desc; /* Indirect descriptor, if any. */
> > + bool map_inter; /* Do dma map internally. */
>
> I prefer a full name. E.g. "dma_map_internal". Eschew abbreviation.
Will fix.
Thanks.
>
>
> > };
> >
> > struct vring_desc_state_packed {
> > @@ -448,7 +449,7 @@ static void vring_unmap_one_split_indirect(const struct
> > vring_virtqueue *vq,
> > }
> >
> > static unsigned int vring_unmap_one_split(const struct vring_virtqueue *vq,
> > - unsigned int i)
> > + unsigned int i, bool map_inter)
> > {
> > struct vring_desc_extra *extra = vq->split.desc_extra;
> > u16 flags;
> > @@ -465,6 +466,9 @@ static unsigned int vring_unmap_one_split(const struct
> > vring_virtqueue *vq,
> > (flags & VRING_DESC_F_WRITE) ?
> > DMA_FROM_DEVICE : DMA_TO_DEVICE);
> > } else {
> > + if (!map_inter)
> > + goto out;
> > +
> > dma_unmap_page(vring_dma_dev(vq),
> > extra[i].addr,
> > extra[i].len,
> > @@ -615,7 +619,7 @@ static inline int virtqueue_add_split(struct virtqueue
> > *_vq,
> > struct scatterlist *sg;
> > struct vring_desc *desc;
> > unsigned int i, n, avail, descs_used, prev;
> > - bool indirect;
> > + bool indirect, map_inter;
> > int head;
> >
> > START_USE(vq);
> > @@ -668,7 +672,8 @@ static inline int virtqueue_add_split(struct virtqueue
> > *_vq,
> > return -ENOSPC;
> > }
> >
> > - if (virtqueue_map_sgs(vq, sgs, total_sg, out_sgs, in_sgs))
> > + map_inter = !sgs[0]->dma_address;
> > + if (map_inter && virtqueue_map_sgs(vq, sgs, total_sg, out_sgs, in_sgs))
> > return -ENOMEM;
> >
> > for (n = 0; n < out_sgs; n++) {
> > @@ -734,6 +739,7 @@ static inline int virtqueue_add_split(struct virtqueue
> > *_vq,
> > vq->split.desc_state[head].indir_desc = desc;
> > else
> > vq->split.desc_state[head].indir_desc = ctx;
> > + vq->split.desc_state[head].map_inter = map_inter;
> >
> > /* Put entry in available array (but don't update avail->idx until they
> > * do sync). */
> > @@ -759,7 +765,8 @@ static inline int virtqueue_add_split(struct virtqueue
> > *_vq,
> > return 0;
> >
> > unmap_release:
> > - virtqueue_unmap_sgs(vq, sgs, total_sg, out_sgs, in_sgs);
> > + if (map_inter)
> > + virtqueue_unmap_sgs(vq, sgs, total_sg, out_sgs, in_sgs);
> >
> > if (indirect)
> > kfree(desc);
> > @@ -804,20 +811,22 @@ static void detach_buf_split(struct vring_virtqueue
> > *vq, unsigned int head,
> > {
> > unsigned int i, j;
> > __virtio16 nextflag = cpu_to_virtio16(vq->vq.vdev, VRING_DESC_F_NEXT);
> > + bool map_inter;
> >
> > /* Clear data ptr. */
> > vq->split.desc_state[head].data = NULL;
> > + map_inter = vq->split.desc_state[head].map_inter;
> >
> > /* Put back on free list: unmap first-level descriptors and find end */
> > i = head;
> >
> > while (vq->split.vring.desc[i].flags & nextflag) {
> > - vring_unmap_one_split(vq, i);
> > + vring_unmap_one_split(vq, i, map_inter);
> > i = vq->split.desc_extra[i].next;
> > vq->vq.num_free++;
> > }
> >
> > - vring_unmap_one_split(vq, i);
> > + vring_unmap_one_split(vq, i, map_inter);
> > vq->split.desc_extra[i].next = vq->free_head;
> > vq->free_head = head;
> >
> > @@ -839,8 +848,10 @@ static void detach_buf_split(struct vring_virtqueue
> > *vq, unsigned int head,
> > VRING_DESC_F_INDIRECT));
> > BUG_ON(len == 0 || len % sizeof(struct vring_desc));
> >
> > - for (j = 0; j < len / sizeof(struct vring_desc); j++)
> > - vring_unmap_one_split_indirect(vq, &indir_desc[j]);
> > + if (map_inter) {
> > + for (j = 0; j < len / sizeof(struct vring_desc); j++)
> > + vring_unmap_one_split_indirect(vq,
> > &indir_desc[j]);
> > + }
> >
> > kfree(indir_desc);
> > vq->split.desc_state[head].indir_desc = NULL;
> > --
> > 2.32.0.3.g01195cf9f
>
_______________________________________________
Virtualization mailing list
[email protected]
https://lists.linuxfoundation.org/mailman/listinfo/virtualization