On Thu, Oct 02, 2025 at 05:46:10PM +0200, Frederic Weisbecker wrote:
> Le Wed, Oct 01, 2025 at 07:48:13AM -0700, Paul E. McKenney a écrit :
> > This commit saves more than 500 lines of RCU code by re-implementing
> > RCU Tasks Trace in terms of SRCU-fast.  Follow-up work will remove
> > more code that does not cause problems by its presence, but that is no
> > longer required.
> > 
> > This variant places smp_mb() in rcu_read_{,un}lock_trace(), which will
> > be removed on common-case architectures in a later commit.
> 
> The changelog doesn't mention what this is ordering :-)

"The ordering that dare not be named"?  ;-)

How about like this for that second paragraph?

        This variant places smp_mb() in rcu_read_{,un}lock_trace(),
        which will be removed on common-case architectures in a
        later commit.  In the meantime, it serves to enforce ordering
        between the underlying srcu_read_{,un}lock_fast() markers and
        the intervening critical section, even on architectures that
        permit attaching tracepoints on regions of code not watched
        by RCU.  Such architectures defeat SRCU-fast's use of implicit
        single-instruction, interrupts-disabled, and atomic-operation
        RCU read-side critical sections, which have no effect when RCU is
        not watching.  The aforementioned later commit will insert these
        smp_mb() calls only on architectures that have not used noinstr to
        prevent attaching tracepoints to code where RCU is not watching.

> > [ paulmck: Apply kernel test robot, Boqun Feng, and Zqiang feedback. ]
> > [ paulmck: Split out Tiny SRCU fixes per Andrii Nakryiko feedback. ]
> > 
> > Signed-off-by: Paul E. McKenney <[email protected]>
> > Tested-by: kernel test robot <[email protected]>
> > Cc: Andrii Nakryiko <[email protected]>
> > Cc: Alexei Starovoitov <[email protected]>
> > Cc: Peter Zijlstra <[email protected]>
> > Cc: <[email protected]>
> > ---
> [...]
> > @@ -50,12 +50,14 @@ static inline void rcu_read_lock_trace(void)
> >  {
> >     struct task_struct *t = current;
> >  
> > -   WRITE_ONCE(t->trc_reader_nesting, READ_ONCE(t->trc_reader_nesting) + 1);
> > -   barrier();
> > -   if (IS_ENABLED(CONFIG_TASKS_TRACE_RCU_READ_MB) &&
> > -       t->trc_reader_special.b.need_mb)
> > -           smp_mb(); // Pairs with update-side barriers
> > -   rcu_lock_acquire(&rcu_trace_lock_map);
> > +   if (t->trc_reader_nesting++) {
> > +           // In case we interrupted a Tasks Trace RCU reader.
> > +           rcu_try_lock_acquire(&rcu_tasks_trace_srcu_struct.dep_map);
> > +           return;
> > +   }
> > +   barrier();  // nesting before scp to protect against interrupt handler.
> > +   t->trc_reader_scp = srcu_read_lock_fast(&rcu_tasks_trace_srcu_struct);
> > +   smp_mb(); // Placeholder for more selective ordering
> 
> Mysterious :-)

Does the reworked commit-log paragraph help clear up this mystery?

> >  }
> >  
> >  /**
> > @@ -69,26 +71,75 @@ static inline void rcu_read_lock_trace(void)
> >   */
> >  static inline void rcu_read_unlock_trace(void)
> >  {
> > -   int nesting;
> > +   struct srcu_ctr __percpu *scp;
> >     struct task_struct *t = current;
> >  
> > -   rcu_lock_release(&rcu_trace_lock_map);
> > -   nesting = READ_ONCE(t->trc_reader_nesting) - 1;
> > -   barrier(); // Critical section before disabling.
> > -   // Disable IPI-based setting of .need_qs.
> > -   WRITE_ONCE(t->trc_reader_nesting, INT_MIN + nesting);
> > -   if (likely(!READ_ONCE(t->trc_reader_special.s)) || nesting) {
> > -           WRITE_ONCE(t->trc_reader_nesting, nesting);
> > -           return;  // We assume shallow reader nesting.
> > -   }
> > -   WARN_ON_ONCE(nesting != 0);
> > -   rcu_read_unlock_trace_special(t);
> > +   smp_mb(); // Placeholder for more selective ordering
> 
> Bizarre :-)

And this bizarreness?  ;-)

> > +   scp = t->trc_reader_scp;
> > +   barrier();  // scp before nesting to protect against interrupt handler.
> 
> What is it protecting against interrupt?

The incrementing of ->trc_reader_nesting vs the fetch of ->trc_reader_scp.

                                                        Thanx, Paul

> > +   if (!--t->trc_reader_nesting)
> > +           srcu_read_unlock_fast(&rcu_tasks_trace_srcu_struct, scp);
> > +   else
> > +           srcu_lock_release(&rcu_tasks_trace_srcu_struct.dep_map);
> > +}
> 
> Thanks (very happy to see all the rest of the code going away!)
> 
> -- 
> Frederic Weisbecker
> SUSE Labs

Reply via email to