On Wed, 2025-09-03 at 11:18 +0100, Tvrtko Ursulin wrote:
> Now that the run queue to scheduler relationship is always 1:1 we can
> embed it (the run queue) directly in the scheduler struct and save on
> some allocation error handling code and such.

Looks reasonable to me.

What I suggest is to do things like general improvements by renaming
variables (see my comments in the previous patch) in a separate
cleanup-patch following this one.

Few comments below

> 
> Signed-off-by: Tvrtko Ursulin <tvrtko.ursu...@igalia.com>
> Cc: Christian König <christian.koe...@amd.com>
> Cc: Danilo Krummrich <d...@kernel.org>
> Cc: Matthew Brost <matthew.br...@intel.com>
> Cc: Philipp Stanner <pha...@kernel.org>
> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c      |  6 ++--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_job.c     |  6 ++--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_job.h     |  5 +++-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h   |  8 ++++--
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c |  8 +++---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c     |  8 +++---
>  drivers/gpu/drm/scheduler/sched_entity.c    | 32 +++++++++------------
>  drivers/gpu/drm/scheduler/sched_fence.c     |  2 +-
>  drivers/gpu/drm/scheduler/sched_internal.h  |  6 ++--
>  drivers/gpu/drm/scheduler/sched_main.c      | 31 ++++----------------
>  drivers/gpu/drm/scheduler/sched_rq.c        | 18 ++++++------
>  include/drm/gpu_scheduler.h                 |  4 +--
>  12 files changed, 58 insertions(+), 76 deletions(-)
> 
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> index 2ac9729e4c86..467e09e5567b 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> @@ -1131,7 +1131,8 @@ static int amdgpu_cs_vm_handling(struct 
> amdgpu_cs_parser *p)
>       if (p->gang_size > 1 && !adev->vm_manager.concurrent_flush) {
>               for (i = 0; i < p->gang_size; ++i) {
>                       struct drm_sched_entity *entity = p->entities[i];
> -                     struct drm_gpu_scheduler *sched = entity->rq->sched;
> +                     struct drm_gpu_scheduler *sched =
> +                             container_of(entity->rq, typeof(*sched), rq);
>                       struct amdgpu_ring *ring = to_amdgpu_ring(sched);
>  
>                       if (amdgpu_vmid_uses_reserved(vm, ring->vm_hub))
> @@ -1262,7 +1263,8 @@ static int amdgpu_cs_sync_rings(struct amdgpu_cs_parser 
> *p)
>                       return r;
>       }

Now that I think of it, are you compiling this for all DRM drivers? Who
knows who might do similar things as amdgpu.

>  
> -     sched = p->gang_leader->base.entity->rq->sched;
> +     sched = container_of(p->gang_leader->base.entity->rq, typeof(*sched),
> +                          rq);
>       while ((fence = amdgpu_sync_get_fence(&p->sync))) {
>               struct drm_sched_fence *s_fence = to_drm_sched_fence(fence);
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> index bc07fd57310c..cdfaf3eb736d 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> @@ -341,7 +341,9 @@ static struct dma_fence *
>  amdgpu_job_prepare_job(struct drm_sched_job *sched_job,
>                     struct drm_sched_entity *s_entity)
>  {
> -     struct amdgpu_ring *ring = to_amdgpu_ring(s_entity->rq->sched);
> +     struct drm_gpu_scheduler *sched =
> +             container_of(s_entity->rq, typeof(*sched), rq);
> +     struct amdgpu_ring *ring = to_amdgpu_ring(sched);
>       struct amdgpu_job *job = to_amdgpu_job(sched_job);
>       struct dma_fence *fence;
>       int r;
> @@ -434,7 +436,7 @@ drm_sched_entity_queue_pop(struct drm_sched_entity 
> *entity)
>  
>  void amdgpu_job_stop_all_jobs_on_sched(struct drm_gpu_scheduler *sched)
>  {
> -     struct drm_sched_rq *rq = sched->rq;
> +     struct drm_sched_rq *rq = &sched->rq;
>       struct drm_sched_entity *s_entity;
>       struct drm_sched_job *s_job;
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.h 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.h
> index 4a6487eb6cb5..9530b5da3adc 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.h
> @@ -102,7 +102,10 @@ struct amdgpu_job {
>  
>  static inline struct amdgpu_ring *amdgpu_job_ring(struct amdgpu_job *job)
>  {
> -     return to_amdgpu_ring(job->base.entity->rq->sched);
> +     struct drm_gpu_scheduler *sched =
> +             container_of(job->base.entity->rq, typeof(*sched), rq);
> +
> +     return to_amdgpu_ring(sched);
>  }
>  
>  int amdgpu_job_alloc(struct amdgpu_device *adev, struct amdgpu_vm *vm,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> index d13e64a69e25..85724ec6aaf8 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h
> @@ -145,6 +145,7 @@ TRACE_EVENT(amdgpu_cs,
>                    struct amdgpu_ib *ib),
>           TP_ARGS(p, job, ib),
>           TP_STRUCT__entry(
> +                          __field(struct drm_gpu_scheduler *, sched)
>                            __field(struct amdgpu_bo_list *, bo_list)
>                            __field(u32, ring)
>                            __field(u32, dw)
> @@ -152,11 +153,14 @@ TRACE_EVENT(amdgpu_cs,
>                            ),
>  
>           TP_fast_assign(
> +                        __entry->sched = container_of(job->base.entity->rq,
> +                                                      
> typeof(*__entry->sched),
> +                                                      rq);
>                          __entry->bo_list = p->bo_list;
> -                        __entry->ring = 
> to_amdgpu_ring(job->base.entity->rq->sched)->idx;
> +                        __entry->ring = to_amdgpu_ring(__entry->sched)->idx;
>                          __entry->dw = ib->length_dw;
>                          __entry->fences = amdgpu_fence_count_emitted(
> -                             to_amdgpu_ring(job->base.entity->rq->sched));
> +                             to_amdgpu_ring(__entry->sched));
>                          ),
>           TP_printk("bo_list=%p, ring=%u, dw=%u, fences=%u",
>                     __entry->bo_list, __entry->ring, __entry->dw,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
> index 36805dcfa159..4ccd2e769799 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm_sdma.c
> @@ -106,13 +106,13 @@ static int amdgpu_vm_sdma_prepare(struct 
> amdgpu_vm_update_params *p,
>  static int amdgpu_vm_sdma_commit(struct amdgpu_vm_update_params *p,
>                                struct dma_fence **fence)
>  {
> +     struct drm_gpu_scheduler *sched =
> +             container_of(p->vm->delayed.rq, typeof(*sched), rq);
> +     struct amdgpu_ring *ring =
> +             container_of(sched, struct amdgpu_ring, sched);
>       struct amdgpu_ib *ib = p->job->ibs;
> -     struct amdgpu_ring *ring;
>       struct dma_fence *f;
>  
> -     ring = container_of(p->vm->delayed.rq->sched, struct amdgpu_ring,
> -                         sched);
> -
>       WARN_ON(ib->length_dw == 0);
>       amdgpu_ring_pad_ib(ring, ib);
>  
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c 
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
> index 1083db8cea2e..be17635ac039 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
> @@ -465,15 +465,15 @@ int amdgpu_xcp_open_device(struct amdgpu_device *adev,
>  void amdgpu_xcp_release_sched(struct amdgpu_device *adev,
>                                 struct amdgpu_ctx_entity *entity)
>  {
> -     struct drm_gpu_scheduler *sched;
> -     struct amdgpu_ring *ring;
> +     struct drm_gpu_scheduler *sched =
> +             container_of(entity->entity.rq, typeof(*sched), rq);
>  
>       if (!adev->xcp_mgr)
>               return;
>  
> -     sched = entity->entity.rq->sched;
>       if (drm_sched_wqueue_ready(sched)) {
> -             ring = to_amdgpu_ring(entity->entity.rq->sched);
> +             struct amdgpu_ring *ring = to_amdgpu_ring(sched);
> +
>               atomic_dec(&adev->xcp_mgr->xcp[ring->xcp_id].ref_cnt);
>       }
>  }
> diff --git a/drivers/gpu/drm/scheduler/sched_entity.c 
> b/drivers/gpu/drm/scheduler/sched_entity.c
> index 6dd30b85925b..ba290143c95d 100644
> --- a/drivers/gpu/drm/scheduler/sched_entity.c
> +++ b/drivers/gpu/drm/scheduler/sched_entity.c
> @@ -114,19 +114,12 @@ int drm_sched_entity_init(struct drm_sched_entity 
> *entity,
>        * is initialized itself.
>        */
>       entity->sched_list = num_sched_list > 1 ? sched_list : NULL;
> +     if (num_sched_list) {
> +             entity->sched_list = num_sched_list > 1 ? sched_list : NULL;
> +             entity->rq = &sched_list[0]->rq;
> +     }
>       RCU_INIT_POINTER(entity->last_scheduled, NULL);
>       RB_CLEAR_NODE(&entity->rb_tree_node);
> -
> -     if (num_sched_list && !sched_list[0]->rq) {
> -             /* Since every entry covered by num_sched_list
> -              * should be non-NULL and therefore we warn drivers
> -              * not to do this and to fix their DRM calling order.
> -              */
> -             pr_warn("%s: called with uninitialized scheduler\n", __func__);
> -     } else if (num_sched_list) {
> -             entity->rq = sched_list[0]->rq;
> -     }
> -
>       init_completion(&entity->entity_idle);
>  
>       /* We start in an idle state. */
> @@ -312,7 +305,7 @@ long drm_sched_entity_flush(struct drm_sched_entity 
> *entity, long timeout)
>       if (!entity->rq)
>               return 0;
>  
> -     sched = entity->rq->sched;
> +     sched = container_of(entity->rq, typeof(*sched), rq);
>       /*
>        * The client will not queue more jobs during this fini - consume
>        * existing queued ones, or discard them on SIGKILL.
> @@ -393,10 +386,12 @@ static void drm_sched_entity_wakeup(struct dma_fence *f,
>  {
>       struct drm_sched_entity *entity =
>               container_of(cb, struct drm_sched_entity, cb);
> +     struct drm_gpu_scheduler *sched =
> +             container_of(entity->rq, typeof(*sched), rq);
>  
>       entity->dependency = NULL;
>       dma_fence_put(f);
> -     drm_sched_wakeup(entity->rq->sched);
> +     drm_sched_wakeup(sched);
>  }
>  
>  /**
> @@ -423,7 +418,8 @@ EXPORT_SYMBOL(drm_sched_entity_set_priority);
>  static bool drm_sched_entity_add_dependency_cb(struct drm_sched_entity 
> *entity,
>                                              struct drm_sched_job *sched_job)
>  {
> -     struct drm_gpu_scheduler *sched = entity->rq->sched;
> +     struct drm_gpu_scheduler *sched =
> +             container_of(entity->rq, typeof(*sched), rq);
>       struct dma_fence *fence = entity->dependency;
>       struct drm_sched_fence *s_fence;
>  
> @@ -557,7 +553,7 @@ void drm_sched_entity_select_rq(struct drm_sched_entity 
> *entity)
>  
>       spin_lock(&entity->lock);
>       sched = drm_sched_pick_best(entity->sched_list, entity->num_sched_list);
> -     rq = sched ? sched->rq : NULL;
> +     rq = sched ? &sched->rq : NULL;
>       if (rq != entity->rq) {
>               drm_sched_rq_remove_entity(entity->rq, entity);
>               entity->rq = rq;
> @@ -580,6 +576,8 @@ void drm_sched_entity_select_rq(struct drm_sched_entity 
> *entity)
>  void drm_sched_entity_push_job(struct drm_sched_job *sched_job)
>  {
>       struct drm_sched_entity *entity = sched_job->entity;
> +     struct drm_gpu_scheduler *sched =
> +             container_of(entity->rq, typeof(*sched), rq);
>       bool first;
>  
>       trace_drm_sched_job_queue(sched_job, entity);
> @@ -591,7 +589,7 @@ void drm_sched_entity_push_job(struct drm_sched_job 
> *sched_job)
>               xa_for_each(&sched_job->dependencies, index, entry)
>                       trace_drm_sched_job_add_dep(sched_job, entry);
>       }
> -     atomic_inc(entity->rq->sched->score);
> +     atomic_inc(sched->score);
>       WRITE_ONCE(entity->last_user, current->group_leader);
>  
>       /*
> @@ -602,8 +600,6 @@ void drm_sched_entity_push_job(struct drm_sched_job 
> *sched_job)
>  
>       /* first job wakes up scheduler */
>       if (first) {
> -             struct drm_gpu_scheduler *sched;
> -
>               sched = drm_sched_rq_add_entity(entity);
>               if (sched)
>                       drm_sched_wakeup(sched);
> diff --git a/drivers/gpu/drm/scheduler/sched_fence.c 
> b/drivers/gpu/drm/scheduler/sched_fence.c
> index 9391d6f0dc01..da4f53a9ca35 100644
> --- a/drivers/gpu/drm/scheduler/sched_fence.c
> +++ b/drivers/gpu/drm/scheduler/sched_fence.c
> @@ -227,7 +227,7 @@ void drm_sched_fence_init(struct drm_sched_fence *fence,
>  {
>       unsigned seq;
>  
> -     fence->sched = entity->rq->sched;
> +     fence->sched = container_of(entity->rq, typeof(*fence->sched), rq);
>       seq = atomic_inc_return(&entity->fence_seq);
>       dma_fence_init(&fence->scheduled, &drm_sched_fence_ops_scheduled,
>                      &fence->lock, entity->fence_context, seq);
> diff --git a/drivers/gpu/drm/scheduler/sched_internal.h 
> b/drivers/gpu/drm/scheduler/sched_internal.h
> index 6e5ed721bb5b..409c9ab7ce8f 100644
> --- a/drivers/gpu/drm/scheduler/sched_internal.h
> +++ b/drivers/gpu/drm/scheduler/sched_internal.h
> @@ -26,11 +26,9 @@ bool drm_sched_can_queue(struct drm_gpu_scheduler *sched,
>                        struct drm_sched_entity *entity);
>  void drm_sched_wakeup(struct drm_gpu_scheduler *sched);
>  
> -void drm_sched_rq_init(struct drm_gpu_scheduler *sched,
> -                    struct drm_sched_rq *rq);
> +void drm_sched_rq_init(struct drm_gpu_scheduler *sched);
>  struct drm_sched_entity *
> -drm_sched_rq_select_entity(struct drm_gpu_scheduler *sched,
> -                        struct drm_sched_rq *rq);
> +drm_sched_rq_select_entity(struct drm_gpu_scheduler *sched);
>  struct drm_gpu_scheduler *
>  drm_sched_rq_add_entity(struct drm_sched_entity *entity);
>  void drm_sched_rq_remove_entity(struct drm_sched_rq *rq,
> diff --git a/drivers/gpu/drm/scheduler/sched_main.c 
> b/drivers/gpu/drm/scheduler/sched_main.c
> index a9079bdb27d3..fe773154cdd4 100644
> --- a/drivers/gpu/drm/scheduler/sched_main.c
> +++ b/drivers/gpu/drm/scheduler/sched_main.c
> @@ -645,7 +645,7 @@ void drm_sched_job_arm(struct drm_sched_job *job)
>  
>       BUG_ON(!entity);
>       drm_sched_entity_select_rq(entity);
> -     sched = entity->rq->sched;
> +     sched = container_of(entity->rq, typeof(*sched), rq);
>  
>       job->sched = sched;
>       job->s_priority = entity->priority;
> @@ -986,7 +986,7 @@ static void drm_sched_run_job_work(struct work_struct *w)
>       int r;
>  
>       /* Find entity with a ready job */
> -     entity = drm_sched_rq_select_entity(sched, sched->rq);
> +     entity = drm_sched_rq_select_entity(sched);
>       if (IS_ERR_OR_NULL(entity))
>               return; /* No more work */
>  
> @@ -1067,15 +1067,6 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, 
> const struct drm_sched_init_
>       sched->score = args->score ? args->score : &sched->_score;
>       sched->dev = args->dev;
>  
> -     if (sched->rq) {
> -             /* Not an error, but warn anyway so drivers can
> -              * fine-tune their DRM calling order, and return all
> -              * is good.
> -              */
> -             dev_warn(sched->dev, "%s: scheduler already initialized!\n", 
> __func__);
> -             return 0;
> -     }
> -
>       if (args->submit_wq) {
>               sched->submit_wq = args->submit_wq;
>               sched->own_submit_wq = false;
> @@ -1087,11 +1078,7 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, 
> const struct drm_sched_init_
>               sched->own_submit_wq = true;
>       }
>  
> -     sched->rq = kmalloc(sizeof(*sched->rq), GFP_KERNEL | __GFP_ZERO);
> -     if (!sched->rq)
> -             goto Out_check_own;
> -
> -     drm_sched_rq_init(sched, sched->rq);
> +     drm_sched_rq_init(sched);
>  
>       init_waitqueue_head(&sched->job_scheduled);
>       INIT_LIST_HEAD(&sched->pending_list);
> @@ -1106,12 +1093,6 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, 
> const struct drm_sched_init_
>  
>       sched->ready = true;
>       return 0;
> -
> -Out_check_own:
> -     if (sched->own_submit_wq)
> -             destroy_workqueue(sched->submit_wq);
> -     dev_err(sched->dev, "%s: Failed to setup GPU scheduler--out of 
> memory\n", __func__);
> -     return -ENOMEM;
>  }
>  EXPORT_SYMBOL(drm_sched_init);
>  
> @@ -1143,7 +1124,7 @@ static void drm_sched_cancel_remaining_jobs(struct 
> drm_gpu_scheduler *sched)
>  void drm_sched_fini(struct drm_gpu_scheduler *sched)
>  {
>  
> -     struct drm_sched_rq *rq = sched->rq;
> +     struct drm_sched_rq *rq = &sched->rq;
>       struct drm_sched_entity *s_entity;
>  
>       drm_sched_wqueue_stop(sched);
> @@ -1185,8 +1166,6 @@ void drm_sched_fini(struct drm_gpu_scheduler *sched)
>       if (sched->own_submit_wq)
>               destroy_workqueue(sched->submit_wq);
>       sched->ready = false;
> -     kfree(sched->rq);
> -     sched->rq = NULL;
>  
>       if (!list_empty(&sched->pending_list))
>               dev_warn(sched->dev, "Tearing down scheduler while jobs are 
> pending!\n");
> @@ -1206,7 +1185,7 @@ void drm_sched_increase_karma(struct drm_sched_job *bad)
>  {
>       struct drm_gpu_scheduler *sched = bad->sched;
>       struct drm_sched_entity *entity, *tmp;
> -     struct drm_sched_rq *rq = sched->rq;
> +     struct drm_sched_rq *rq = &sched->rq;
>  
>       /* don't change @bad's karma if it's from KERNEL RQ,
>        * because sometimes GPU hang would cause kernel jobs (like VM updating 
> jobs)
> diff --git a/drivers/gpu/drm/scheduler/sched_rq.c 
> b/drivers/gpu/drm/scheduler/sched_rq.c
> index a9bc105221bf..6088434a4ea4 100644
> --- a/drivers/gpu/drm/scheduler/sched_rq.c
> +++ b/drivers/gpu/drm/scheduler/sched_rq.c
> @@ -72,17 +72,16 @@ static void drm_sched_rq_update_tree_locked(struct 
> drm_sched_entity *entity,
>   * drm_sched_rq_init - initialize a given run queue struct
>   *
>   * @sched: scheduler instance to associate with this run queue
> - * @rq: scheduler run queue
>   *
>   * Initializes a scheduler runqueue.
>   */
> -void drm_sched_rq_init(struct drm_gpu_scheduler *sched,
> -                    struct drm_sched_rq *rq)
> +void drm_sched_rq_init(struct drm_gpu_scheduler *sched)
>  {
> +     struct drm_sched_rq *rq = &sched->rq;
> +
>       spin_lock_init(&rq->lock);
>       INIT_LIST_HEAD(&rq->entities);
>       rq->rb_tree_root = RB_ROOT_CACHED;
> -     rq->sched = sched;
>       rq->head_prio = -1;
>  }
>  
> @@ -229,8 +228,8 @@ drm_sched_rq_add_entity(struct drm_sched_entity *entity)
>       }
>  
>       rq = entity->rq;
> +     sched = container_of(rq, typeof(*sched), rq);
>       spin_lock(&rq->lock);
> -     sched = rq->sched;
>  
>       if (list_empty(&entity->list)) {
>               atomic_inc(sched->score);
> @@ -258,6 +257,8 @@ drm_sched_rq_add_entity(struct drm_sched_entity *entity)
>  void drm_sched_rq_remove_entity(struct drm_sched_rq *rq,
>                               struct drm_sched_entity *entity)
>  {
> +     struct drm_gpu_scheduler *sched = container_of(rq, typeof(*sched), rq);
> +
>       lockdep_assert_held(&entity->lock);
>  
>       if (list_empty(&entity->list))
> @@ -265,7 +266,7 @@ void drm_sched_rq_remove_entity(struct drm_sched_rq *rq,
>  
>       spin_lock(&rq->lock);
>  
> -     atomic_dec(rq->sched->score);
> +     atomic_dec(sched->score);
>       list_del_init(&entity->list);
>  
>       drm_sched_rq_remove_tree_locked(entity, rq);
> @@ -313,7 +314,6 @@ void drm_sched_rq_pop_entity(struct drm_sched_entity 
> *entity)
>   * drm_sched_rq_select_entity - Select an entity which provides a job to run
>   *
>   * @sched: the gpu scheduler
> - * @rq: scheduler run queue to check.
>   *
>   * Find oldest waiting ready entity.
>   *
> @@ -322,9 +322,9 @@ void drm_sched_rq_pop_entity(struct drm_sched_entity 
> *entity)
>   * its job; return NULL, if no ready entity was found.
>   */
>  struct drm_sched_entity *
> -drm_sched_rq_select_entity(struct drm_gpu_scheduler *sched,
> -                        struct drm_sched_rq *rq)
> +drm_sched_rq_select_entity(struct drm_gpu_scheduler *sched)
>  {
> +     struct drm_sched_rq *rq = &sched->rq;
>       struct rb_node *rb;
>  
>       spin_lock(&rq->lock);
> diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h
> index c9b75a05d05c..7fbcd121a6d3 100644
> --- a/include/drm/gpu_scheduler.h
> +++ b/include/drm/gpu_scheduler.h
> @@ -256,8 +256,6 @@ struct drm_sched_entity {
>   * the next entity to emit commands from.
>   */
>  struct drm_sched_rq {
> -     struct drm_gpu_scheduler        *sched;
> -

Docu needs to be updated


--

So much from my side for now. I guess we'll be discussing a bit on the
other threads and on XDC. But overall, this is a very nice series that
improves things greatly, and from my POV we can transition to a v1;
maybe after brainstorming with the other guys next week a bit.


Regards,
P.

>       spinlock_t                      lock;
>       /* Following members are protected by the @lock: */
>       struct list_head                entities;
> @@ -581,7 +579,7 @@ struct drm_gpu_scheduler {
>       atomic_t                        credit_count;
>       long                            timeout;
>       const char                      *name;
> -     struct drm_sched_rq             *rq;
> +     struct drm_sched_rq             rq;
>       wait_queue_head_t               job_scheduled;
>       atomic64_t                      job_id_count;
>       struct workqueue_struct         *submit_wq;

Reply via email to