On Tue, Jun 13, 2017 at 02:49:47PM -0400, Rob Clark wrote:
> It means we have to do a list traversal where we once had an index into
> a table.  But the list will normally have one or two entries.
> 
> Signed-off-by: Rob Clark <[email protected]>

I dig the rename - makes more sense.

Acked-by: Jordan Crouse <[email protected]>

> ---
>  drivers/gpu/drm/msm/msm_gem.c | 138 
> +++++++++++++++++++++++++++++-------------
>  drivers/gpu/drm/msm/msm_gem.h |   4 +-
>  2 files changed, 99 insertions(+), 43 deletions(-)
> 
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index 754432c..410368f 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -283,21 +283,59 @@ uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj)
>       return offset;
>  }
>  
> +static struct msm_gem_vma *add_vma(struct drm_gem_object *obj,
> +             struct msm_gem_address_space *aspace)
> +{
> +     struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +     struct msm_gem_vma *vma;
> +
> +     vma = kzalloc(sizeof(*vma), GFP_KERNEL);
> +     if (!vma)
> +             return ERR_PTR(-ENOMEM);
> +
> +     vma->aspace = aspace;
> +
> +     list_add_tail(&vma->list, &msm_obj->vmas);
> +
> +     return vma;
> +}
> +
> +static struct msm_gem_vma *lookup_vma(struct drm_gem_object *obj,
> +             struct msm_gem_address_space *aspace)
> +{
> +     struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +     struct msm_gem_vma *vma;
> +
> +     WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
> +
> +     list_for_each_entry(vma, &msm_obj->vmas, list) {
> +             if (vma->aspace == aspace)
> +                     return vma;
> +     }
> +
> +     return NULL;
> +}
> +
> +static void del_vma(struct msm_gem_vma *vma)
> +{
> +     if (!vma)
> +             return;
> +
> +     list_del(&vma->list);
> +     kfree(vma);
> +}
> +
>  static void
>  put_iova(struct drm_gem_object *obj)
>  {
> -     struct drm_device *dev = obj->dev;
> -     struct msm_drm_private *priv = obj->dev->dev_private;
>       struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -     int id;
> +     struct msm_gem_vma *vma, *tmp;
>  
> -     WARN_ON(!mutex_is_locked(&dev->struct_mutex));
> +     WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> -     for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) {
> -             if (!priv->aspace[id])
> -                     continue;
> -             msm_gem_unmap_vma(priv->aspace[id],
> -                             &msm_obj->domain[id], msm_obj->sgt);
> +     list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) {
> +             msm_gem_unmap_vma(vma->aspace, vma, msm_obj->sgt);
> +             del_vma(vma);
>       }
>  }
>  
> @@ -312,24 +350,37 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj,
>               struct msm_gem_address_space *aspace, uint64_t *iova)
>  {
>       struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -     int id = aspace ? aspace->id : 0;
> +     struct msm_gem_vma *vma;
>       int ret = 0;
>  
>       WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> -     if (!msm_obj->domain[id].iova) {
> -             struct msm_drm_private *priv = obj->dev->dev_private;
> -             struct page **pages = get_pages(obj);
> +     vma = lookup_vma(obj, aspace);
>  
> -             if (IS_ERR(pages))
> -                     return PTR_ERR(pages);
> +     if (!vma) {
> +             struct page **pages;
> +
> +             vma = add_vma(obj, aspace);
> +             if (IS_ERR(vma))
> +                     return PTR_ERR(vma);
> +
> +             pages = get_pages(obj);
> +             if (IS_ERR(pages)) {
> +                     ret = PTR_ERR(pages);
> +                     goto fail;
> +             }
>  
> -             ret = msm_gem_map_vma(priv->aspace[id], &msm_obj->domain[id],
> -                             msm_obj->sgt, obj->size >> PAGE_SHIFT);
> +             ret = msm_gem_map_vma(aspace, vma, msm_obj->sgt,
> +                             obj->size >> PAGE_SHIFT);
> +             if (ret)
> +                     goto fail;
>       }
>  
> -     if (!ret)
> -             *iova = msm_obj->domain[id].iova;
> +     *iova = vma->iova;
> +     return 0;
> +
> +fail:
> +     del_vma(vma);
>  
>       return ret;
>  }
> @@ -338,22 +389,12 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj,
>  int msm_gem_get_iova(struct drm_gem_object *obj,
>               struct msm_gem_address_space *aspace, uint64_t *iova)
>  {
> -     struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -     int id = aspace ? aspace->id : 0;
>       int ret;
>  
> -     /* this is safe right now because we don't unmap until the
> -      * bo is deleted:
> -      */
> -     if (msm_obj->domain[id].iova) {
> -             might_lock(&obj->dev->struct_mutex);
> -             *iova = msm_obj->domain[id].iova;
> -             return 0;
> -     }
> -
>       mutex_lock(&obj->dev->struct_mutex);
>       ret = msm_gem_get_iova_locked(obj, aspace, iova);
>       mutex_unlock(&obj->dev->struct_mutex);
> +
>       return ret;
>  }
>  
> @@ -363,10 +404,14 @@ int msm_gem_get_iova(struct drm_gem_object *obj,
>  uint64_t msm_gem_iova(struct drm_gem_object *obj,
>               struct msm_gem_address_space *aspace)
>  {
> -     struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -     int id = aspace ? aspace->id : 0;
> -     WARN_ON(!msm_obj->domain[id].iova);
> -     return msm_obj->domain[id].iova;
> +     struct msm_gem_vma *vma;
> +
> +     mutex_lock(&obj->dev->struct_mutex);
> +     vma = lookup_vma(obj, aspace);
> +     mutex_unlock(&obj->dev->struct_mutex);
> +     WARN_ON(!vma);
> +
> +     return vma ? vma->iova : 0;
>  }
>  
>  void msm_gem_put_iova(struct drm_gem_object *obj,
> @@ -624,11 +669,10 @@ void msm_gem_describe(struct drm_gem_object *obj, 
> struct seq_file *m)
>       struct msm_gem_object *msm_obj = to_msm_bo(obj);
>       struct reservation_object *robj = msm_obj->resv;
>       struct reservation_object_list *fobj;
> -     struct msm_drm_private *priv = obj->dev->dev_private;
>       struct dma_fence *fence;
> +     struct msm_gem_vma *vma;
>       uint64_t off = drm_vma_node_start(&obj->vma_node);
>       const char *madv;
> -     unsigned id;
>  
>       WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> @@ -650,8 +694,9 @@ void msm_gem_describe(struct drm_gem_object *obj, struct 
> seq_file *m)
>                       obj->name, kref_read(&obj->refcount),
>                       off, msm_obj->vaddr);
>  
> -     for (id = 0; id < priv->num_aspaces; id++)
> -             seq_printf(m, " %08llx", msm_obj->domain[id].iova);
> +     /* FIXME: we need to print the address space here too */
> +     list_for_each_entry(vma, &msm_obj->vmas, list)
> +             seq_printf(m, " %08llx", vma->iova);
>  
>       seq_printf(m, " %zu%s\n", obj->size, madv);
>  
> @@ -788,6 +833,8 @@ static int msm_gem_new_impl(struct drm_device *dev,
>       }
>  
>       INIT_LIST_HEAD(&msm_obj->submit_entry);
> +     INIT_LIST_HEAD(&msm_obj->vmas);
> +
>       list_add_tail(&msm_obj->mm_list, &priv->inactive_list);
>  
>       *obj = &msm_obj->base;
> @@ -826,19 +873,26 @@ struct drm_gem_object *msm_gem_new(struct drm_device 
> *dev,
>               goto fail;
>  
>       if (use_vram) {
> -             struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +             struct msm_gem_vma *vma;
>               struct page **pages;
>  
> -             msm_obj->vram_node = &msm_obj->domain[0].node;
> +             vma = add_vma(obj, NULL);
> +             if (IS_ERR(vma)) {
> +                     ret = PTR_ERR(vma);
> +                     goto fail;
> +             }
> +
> +             to_msm_bo(obj)->vram_node = &vma->node;
> +
>               drm_gem_private_object_init(dev, obj, size);
>  
> -             msm_obj->pages = get_pages(obj);
>               pages = get_pages(obj);
>               if (IS_ERR(pages)) {
>                       ret = PTR_ERR(pages);
>                       goto fail;
>               }
> -             msm_obj->domain[0].iova = physaddr(obj);
> +
> +             vma->iova = physaddr(obj);
>       } else {
>               ret = drm_gem_object_init(dev, obj, size);
>               if (ret)
> diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
> index 4b4b352..ff468da 100644
> --- a/drivers/gpu/drm/msm/msm_gem.h
> +++ b/drivers/gpu/drm/msm/msm_gem.h
> @@ -39,6 +39,8 @@ struct msm_gem_address_space {
>  struct msm_gem_vma {
>       struct drm_mm_node node;
>       uint64_t iova;
> +     struct msm_gem_address_space *aspace;
> +     struct list_head list;    /* node in msm_gem_object::vmas */
>  };
>  
>  struct msm_gem_object {
> @@ -78,7 +80,7 @@ struct msm_gem_object {
>       struct sg_table *sgt;
>       void *vaddr;
>  
> -     struct msm_gem_vma domain[NUM_DOMAINS];
> +     struct list_head vmas;    /* list of msm_gem_vma */
>  
>       /* normally (resv == &_resv) except for imported bo's */
>       struct reservation_object *resv;
> -- 
> 2.9.4
> 

-- 
The Qualcomm Innovation Center, Inc. is a member of Code Aurora Forum,
a Linux Foundation Collaborative Project
_______________________________________________
dri-devel mailing list
[email protected]
https://lists.freedesktop.org/mailman/listinfo/dri-devel

Reply via email to