On Tue, Mar 3, 2026 at 8:28 PM Zhang, Jesse(Jie) <[email protected]> wrote:
>
> [AMD Official Use Only - AMD Internal Distribution Only]
>
> > -----Original Message-----
> > From: amd-gfx <[email protected]> On Behalf Of Alex
> > Deucher
> > Sent: Tuesday, March 3, 2026 11:23 PM
> > To: Deucher, Alexander <[email protected]>
> > Cc: [email protected]
> > Subject: Re: [PATCH] drm/amdgpu/gfx12.1: add support for disable_kq
> >
> > Ping?
> >
> > Alex
> >
> > On Wed, Jan 28, 2026 at 3:12 PM Alex Deucher <[email protected]>
> > wrote:
> > >
> > > Plumb in support for disabling kernel queues and make it the default.
> > > For testing, kernel queues can be re-enabled by setting
> > > amdgpu.user_queue=0
> > >
> > > Signed-off-by: Alex Deucher <[email protected]>
> > > ---
> > >  drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c | 109
> > > +++++++++++++++++++------
> > >  1 file changed, 82 insertions(+), 27 deletions(-)
> > >
> > > diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > > b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > > index 08ae50a6313f3..f93ee275ce398 100644
> > > --- a/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > > +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
> > > @@ -1155,11 +1155,13 @@ static int gfx_v12_1_sw_init(struct 
> > > amdgpu_ip_block
> > *ip_block)
> > >                 break;
> > >         }
> > >
> > > -       /* recalculate compute rings to use based on hardware 
> > > configuration */
> > > -       num_compute_rings = (adev->gfx.mec.num_pipe_per_mec *
> > > -                            adev->gfx.mec.num_queue_per_pipe) / 2;
> > > -       adev->gfx.num_compute_rings = min(adev->gfx.num_compute_rings,
> > > -                                         num_compute_rings);
> > > +       if (adev->gfx.num_compute_rings) {
> > > +               /* recalculate compute rings to use based on hardware 
> > > configuration */
> > > +               num_compute_rings = (adev->gfx.mec.num_pipe_per_mec *
> > > +                                    adev->gfx.mec.num_queue_per_pipe) / 
> > > 2;
> > > +               adev->gfx.num_compute_rings = 
> > > min(adev->gfx.num_compute_rings,
> > > +                                                 num_compute_rings);
> > > +       }
> > >
> > >         num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> > >
> > > @@ -2794,6 +2796,36 @@ static void gfx_v12_1_xcc_fini(struct
> > amdgpu_device *adev,
> > >         gfx_v12_1_xcc_enable_gui_idle_interrupt(adev, false, xcc_id);
> > > }
> > >
> > > +static int gfx_v12_1_set_userq_eop_interrupts(struct amdgpu_device *adev,
> > > +                                             bool enable) {
> > > +       unsigned int irq_type;
> > > +       int m, p, r, x, num_xcc;
> > > +
> > > +       if (adev->gfx.disable_kq) {
> > > +               num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> > > +               for (x = 0; x < num_xcc; x++) {
> > > +                       for (m = 0; m < adev->gfx.mec.num_mec; ++m) {
> > > +                               for (p = 0; p < 
> > > adev->gfx.mec.num_pipe_per_mec; p++) {
> > > +                                       irq_type =
> > AMDGPU_CP_IRQ_COMPUTE_MEC1_PIPE0_EOP
> > > +                                               + (m * 
> > > adev->gfx.mec.num_pipe_per_mec)
> > > +                                               + p;
> > > +                                       if (enable)
> > > +                                               r = amdgpu_irq_get(adev, 
> > > &adev->gfx.eop_irq,
> > > +                                                                  
> > > irq_type);
> > > +                                       else
> > > +                                               r = amdgpu_irq_put(adev, 
> > > &adev->gfx.eop_irq,
> > > +                                                                  
> > > irq_type);
> > > +                                       if (r)
> > > +                                               return r;
> > > +                               }
> > > +                       }
> > > +               }
> > > +       }
> > > +
> > > +       return 0;
> > > +}
> > > +
> > >  static int gfx_v12_1_hw_fini(struct amdgpu_ip_block *ip_block)  {
> > >         struct amdgpu_device *adev = ip_block->adev; @@ -2801,6
> > > +2833,7 @@ static int gfx_v12_1_hw_fini(struct amdgpu_ip_block
> > > *ip_block)
> > >
> > >         amdgpu_irq_put(adev, &adev->gfx.priv_reg_irq, 0);
> > >         amdgpu_irq_put(adev, &adev->gfx.priv_inst_irq, 0);
> > > +       gfx_v12_1_set_userq_eop_interrupts(adev, false);
> > >
> > >         num_xcc = NUM_XCC(adev->gfx.xcc_mask);
> > >         for (i = 0; i < num_xcc; i++) { @@ -2868,10 +2901,26 @@ static
> > > int gfx_v12_1_early_init(struct amdgpu_ip_block *ip_block)  {
> > >         struct amdgpu_device *adev = ip_block->adev;
> > >
> > > +
> > > +       switch (amdgpu_user_queue) {
> > > +       case -1:
> > > +       default:
> > > +               adev->gfx.disable_kq = true;
> > > +               adev->gfx.disable_uq = true;
> Should we keep it consistent with the gfx11 settings (gfx_v11_0_early_init)?

There's no need for kernel queues or KGD user queues in this case.

Alex

> Except this question, Reviewed-by: Jesse Zhang <[email protected]>
> > > +               break;
> > > +       case 0:
> > > +               adev->gfx.disable_kq = false;
> > > +               adev->gfx.disable_uq = true;
> > > +               break;
> > > +       }
> > > +
> > >         adev->gfx.funcs = &gfx_v12_1_gfx_funcs;
> > >
> > > -       adev->gfx.num_compute_rings = min(amdgpu_gfx_get_num_kcq(adev),
> > > -                                         AMDGPU_MAX_COMPUTE_RINGS);
> > > +       if (adev->gfx.disable_kq)
> > > +               adev->gfx.num_compute_rings = 0;
> > > +       else
> > > +               adev->gfx.num_compute_rings =
> > min(amdgpu_gfx_get_num_kcq(adev),
> > > +
> > > + AMDGPU_MAX_COMPUTE_RINGS);
> > >
> > >         gfx_v12_1_set_kiq_pm4_funcs(adev);
> > >         gfx_v12_1_set_ring_funcs(adev); @@ -2898,6 +2947,10 @@ static
> > > int gfx_v12_1_late_init(struct amdgpu_ip_block *ip_block)
> > >         if (r)
> > >                 return r;
> > >
> > > +       r = gfx_v12_1_set_userq_eop_interrupts(adev, true);
> > > +       if (r)
> > > +               return r;
> > > +
> > >         return 0;
> > >  }
> > >
> > > @@ -3719,29 +3772,31 @@ static void gfx_v12_1_handle_priv_fault(struct
> > amdgpu_device *adev,
> > >         if (xcc_id == -EINVAL)
> > >                 return;
> > >
> > > -       switch (me_id) {
> > > -       case 0:
> > > -               for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
> > > -                       ring = &adev->gfx.gfx_ring[i];
> > > -                       /* we only enabled 1 gfx queue per pipe for now */
> > > -                       if (ring->me == me_id && ring->pipe == pipe_id)
> > > -                               drm_sched_fault(&ring->sched);
> > > -               }
> > > -               break;
> > > -       case 1:
> > > -       case 2:
> > > -               for (i = 0; i < adev->gfx.num_compute_rings; i++) {
> > > -                       ring = &adev->gfx.compute_ring
> > > +       if (!adev->gfx.disable_kq) {
> > > +               switch (me_id) {
> > > +               case 0:
> > > +                       for (i = 0; i < adev->gfx.num_gfx_rings; i++) {
> > > +                               ring = &adev->gfx.gfx_ring[i];
> > > +                               /* we only enabled 1 gfx queue per pipe 
> > > for now */
> > > +                               if (ring->me == me_id && ring->pipe == 
> > > pipe_id)
> > > +                                       drm_sched_fault(&ring->sched);
> > > +                       }
> > > +                       break;
> > > +               case 1:
> > > +               case 2:
> > > +                       for (i = 0; i < adev->gfx.num_compute_rings; i++) 
> > > {
> > > +                               ring = &adev->gfx.compute_ring
> > >                                         [i +
> > >                                          xcc_id * 
> > > adev->gfx.num_compute_rings];
> > > -                       if (ring->me == me_id && ring->pipe == pipe_id &&
> > > -                           ring->queue == queue_id)
> > > -                               drm_sched_fault(&ring->sched);
> > > +                               if (ring->me == me_id && ring->pipe == 
> > > pipe_id &&
> > > +                                   ring->queue == queue_id)
> > > +                                       drm_sched_fault(&ring->sched);
> > > +                       }
> > > +                       break;
> > > +               default:
> > > +                       BUG();
> > > +                       break;
> > >                 }
> > > -               break;
> > > -       default:
> > > -               BUG();
> > > -               break;
> > >         }
> > >  }
> > >
> > > --
> > > 2.52.0
> > >

Reply via email to