On Tue, Mar 3, 2026 at 10:19 PM Lazar, Lijo <[email protected]> wrote:
>
>
>
> On 29-Jan-26 1:23 AM, Alex Deucher wrote:
> > Plumb in support for disabling kernel queues and make it
> > the default.  For testing, kernel queues can be re-enabled
> > by setting amdgpu.user_queue=0
> >
> > Signed-off-by: Alex Deucher <[email protected]>
> > ---
> >   drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c | 109 +++++++++++++++++++------
> >   1 file changed, 82 insertions(+), 27 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c 
> > b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > index 08ae50a6313f3..f93ee275ce398 100644
> > --- a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > @@ -1155,11 +1155,13 @@ static int gfx_v12_1_sw_init(struct amdgpu_ip_block 
> > *ip_block)
> >               break;
> >       }
> >
> > -     /* recalculate compute rings to use based on hardware configuration */
> > -     num_compute_rings = (adev->gfx.mec.num_pipe_per_mec *
> > -                          adev->gfx.mec.num_queue_per_pipe) / 2;
> > -     adev->gfx.num_compute_rings = min(adev->gfx.num_compute_rings,
> > -                                       num_compute_rings);
> > +     if (adev->gfx.num_compute_rings) {
> > +             /* recalculate compute rings to use based on hardware 
> > configuration */
> > +             num_compute_rings = (adev->gfx.mec.num_pipe_per_mec *
> > +                                  adev->gfx.mec.num_queue_per_pipe) / 2;
> > +             adev->gfx.num_compute_rings = min(adev->gfx.num_compute_rings,
> > +                                               num_compute_rings);
> > +     }
> >
> >       num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> >
> > @@ -2794,6 +2796,36 @@ static void gfx_v12_1_xcc_fini(struct amdgpu_device 
> > *adev,
> >       gfx_v12_1_xcc_enable_gui_idle_interrupt(adev, false, xcc_id);
> >   }
> >
> > +static int gfx_v12_1_set_userq_eop_interrupts(struct amdgpu_device *adev,
> > +                                           bool enable)
> > +{
> > +     unsigned int irq_type;
> > +     int m, p, r, x, num_xcc;
> > +
> > +     if (adev->gfx.disable_kq) {
> > +             num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> > +             for (x = 0; x < num_xcc; x++) {
> > +                     for (m = 0; m < adev->gfx.mec.num_mec; ++m) {
> > +                             for (p = 0; p < 
> > adev->gfx.mec.num_pipe_per_mec; p++) {
> > +                                     irq_type = 
> > AMDGPU_CP_IRQ_COMPUTE_MEC1_PIPE0_EOP
> > +                                             + (m * 
> > adev->gfx.mec.num_pipe_per_mec)
> > +                                             + p;
>
> if x is not involved in type calculation, I guess iteration over x is
> not required. Only the handler for the interrupt type needs to be
> enabled once.

will fix.

>
> > +                                     if (enable)
> > +                                             r = amdgpu_irq_get(adev, 
> > &adev->gfx.eop_irq,
> > +                                                                irq_type);
> > +                                     else
> > +                                             r = amdgpu_irq_put(adev, 
> > &adev->gfx.eop_irq,
> > +                                                                irq_type);
> > +                                     if (r)
> > +                                             return r;
> > +                             }
> > +                     }
> > +             }
> > +     }
> > +
> > +     return 0;
> > +}
> > +
> >   static int gfx_v12_1_hw_fini(struct amdgpu_ip_block *ip_block)
> >   {
> >       struct amdgpu_device *adev = ip_block->adev;
> > @@ -2801,6 +2833,7 @@ static int gfx_v12_1_hw_fini(struct amdgpu_ip_block 
> > *ip_block)
> >
> >       amdgpu_irq_put(adev, &adev->gfx.priv_reg_irq, 0);
> >       amdgpu_irq_put(adev, &adev->gfx.priv_inst_irq, 0);
> > +     gfx_v12_1_set_userq_eop_interrupts(adev, false);
> >
> >       num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> >       for (i = 0; i < num_xcc; i++) {
> > @@ -2868,10 +2901,26 @@ static int gfx_v12_1_early_init(struct 
> > amdgpu_ip_block *ip_block)
> >   {
> >       struct amdgpu_device *adev = ip_block->adev;
> >
> > +
> > +     switch (amdgpu_user_queue) {
> > +     case -1:
> > +     default:
> > +             adev->gfx.disable_kq = true;
> > +             adev->gfx.disable_uq = true;
> > +             break;
> > +     case 0:
> > +             adev->gfx.disable_kq = false;
> > +             adev->gfx.disable_uq = true;
> > +             break;
> > +     }
> > +
> >       adev->gfx.funcs = &gfx_v12_1_gfx_funcs;
> >
> > -     adev->gfx.num_compute_rings = min(amdgpu_gfx_get_num_kcq(adev),
> > -                                       AMDGPU_MAX_COMPUTE_RINGS);
> > +     if (adev->gfx.disable_kq)
> > +             adev->gfx.num_compute_rings = 0;
> > +     else
> > +             adev->gfx.num_compute_rings = 
> > min(amdgpu_gfx_get_num_kcq(adev),
> > +                                               AMDGPU_MAX_COMPUTE_RINGS);
> >
> >       gfx_v12_1_set_kiq_pm4_funcs(adev);
> >       gfx_v12_1_set_ring_funcs(adev);
> > @@ -2898,6 +2947,10 @@ static int gfx_v12_1_late_init(struct 
> > amdgpu_ip_block *ip_block)
> >       if (r)
> >               return r;
> >
> > +     r = gfx_v12_1_set_userq_eop_interrupts(adev, true);
> > +     if (r)
> > +             return r;
> > +
> >       return 0;
> >   }
> >
> > @@ -3719,29 +3772,31 @@ static void gfx_v12_1_handle_priv_fault(struct 
> > amdgpu_device *adev,
> >       if (xcc_id == -EINVAL)
> >               return;
> >
> > -     switch (me_id) {
> > -     case 0:
> > -             for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
> > -                     ring = &adev->gfx.gfx_ring[i];
> > -                     /* we only enabled 1 gfx queue per pipe for now */
> > -                     if (ring->me == me_id && ring->pipe == pipe_id)
> > -                             drm_sched_fault(&ring->sched);
> > -             }
> > -             break;
> > -     case 1:
> > -     case 2:
> > -             for (i = 0; i < adev->gfx.num_compute_rings; i++) {
> > -                     ring = &adev->gfx.compute_ring
> > +     if (!adev->gfx.disable_kq) {
>
> If no handled here, is this unexpected or handled somewhere else?

If kernel queues are not enabled, a kernel queue would never be
responsible so nothing would match here anyway.  Support for user
queues still needs to be hooked for these faults in general.

>
> > +             switch (me_id) {
> > +             case 0:
> > +                     for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
> > +                             ring = &adev->gfx.gfx_ring[i];
> > +                             /* we only enabled 1 gfx queue per pipe for 
> > now */
> > +                             if (ring->me == me_id && ring->pipe == 
> > pipe_id)
> > +                                     drm_sched_fault(&ring->sched);
> > +                     }
>
> Not related to this patch, but this code looks redundant.

I can clean this up in a separate patch.

Alex

>
> Thanks,
> Lijo
>
> > +                     break;
> > +             case 1:
> > +             case 2:
> > +                     for (i = 0; i < adev->gfx.num_compute_rings; i++) {
> > +                             ring = &adev->gfx.compute_ring
> >                                       [i +
> >                                        xcc_id * 
> > adev->gfx.num_compute_rings];
> > -                     if (ring->me == me_id && ring->pipe == pipe_id &&
> > -                         ring->queue == queue_id)
> > -                             drm_sched_fault(&ring->sched);
> > +                             if (ring->me == me_id && ring->pipe == 
> > pipe_id &&
> > +                                 ring->queue == queue_id)
> > +                                     drm_sched_fault(&ring->sched);
> > +                     }
> > +                     break;
> > +             default:
> > +                     BUG();
> > +                     break;
> >               }
> > -             break;
> > -     default:
> > -             BUG();
> > -             break;
> >       }
> >   }
> >
>

Reply via email to