On Thu, Nov 20, 2025 at 04:09:39PM +0100, Marco Elver wrote:
> Improve the existing annotations to properly support Clang's context
> analysis.
> 
> The old annotations distinguished between RCU, RCU_BH, and RCU_SCHED;
> however, to more easily be able to express that "hold the RCU read lock"
> without caring if the normal, _bh(), or _sched() variant was used we'd
> have to remove the distinction of the latter variants: change the _bh()
> and _sched() variants to also acquire "RCU".
> 
> When (and if) we introduce context guards to denote more generally that
> "IRQ", "BH", "PREEMPT" contexts are disabled, it would make sense to
> acquire these instead of RCU_BH and RCU_SCHED respectively.
> 
> The above change also simplified introducing __guarded_by support, where
> only the "RCU" context guard needs to be held: introduce __rcu_guarded,
> where Clang's context analysis warns if a pointer is dereferenced
> without any of the RCU locks held, or updated without the appropriate
> helpers.
> 
> The primitives rcu_assign_pointer() and friends are wrapped with
> context_unsafe(), which enforces using them to update RCU-protected
> pointers marked with __rcu_guarded.
> 
> Signed-off-by: Marco Elver <[email protected]>

Good reminder!  I had lost track of this series.

My big questions here are:

o       What about RCU readers using (say) preempt_disable() instead
        of rcu_read_lock_sched()?

o       What about RCU readers using local_bh_disable() instead of
        rcu_read_lock_sched()?

And keeping in mind that such readers might start in assembly language.

One reasonable approach is to require such readers to use something like
rcu_dereference_all() or rcu_dereference_all_check(), which could then
have special dispensation to instead rely on run-time checks.

Another more powerful approach would be to make this facility also
track preemption, interrupt, NMI, and BH contexts.

Either way could be a significant improvement over what we have now.

Thoughts?

                                                        Thanx, Paul

> ---
> v3:
> * Properly support reentrancy via new compiler support.
> 
> v2:
> * Reword commit message and point out reentrancy caveat.
> ---
>  Documentation/dev-tools/context-analysis.rst |  2 +-
>  include/linux/rcupdate.h                     | 77 ++++++++++++------
>  lib/test_context-analysis.c                  | 85 ++++++++++++++++++++
>  3 files changed, 139 insertions(+), 25 deletions(-)
> 
> diff --git a/Documentation/dev-tools/context-analysis.rst 
> b/Documentation/dev-tools/context-analysis.rst
> index a3d925ce2df4..05164804a92a 100644
> --- a/Documentation/dev-tools/context-analysis.rst
> +++ b/Documentation/dev-tools/context-analysis.rst
> @@ -81,7 +81,7 @@ Supported Kernel Primitives
>  
>  Currently the following synchronization primitives are supported:
>  `raw_spinlock_t`, `spinlock_t`, `rwlock_t`, `mutex`, `seqlock_t`,
> -`bit_spinlock`.
> +`bit_spinlock`, RCU.
>  
>  For context guards with an initialization function (e.g., 
> `spin_lock_init()`),
>  calling this function before initializing any guarded members or globals
> diff --git a/include/linux/rcupdate.h b/include/linux/rcupdate.h
> index c5b30054cd01..5cddb9019a99 100644
> --- a/include/linux/rcupdate.h
> +++ b/include/linux/rcupdate.h
> @@ -31,6 +31,16 @@
>  #include <asm/processor.h>
>  #include <linux/context_tracking_irq.h>
>  
> +token_context_guard(RCU, __reentrant_ctx_guard);
> +token_context_guard_instance(RCU, RCU_SCHED);
> +token_context_guard_instance(RCU, RCU_BH);
> +
> +/*
> + * A convenience macro that can be used for RCU-protected globals or struct
> + * members; adds type qualifier __rcu, and also enforces __guarded_by(RCU).
> + */
> +#define __rcu_guarded __rcu __guarded_by(RCU)
> +
>  #define ULONG_CMP_GE(a, b)   (ULONG_MAX / 2 >= (a) - (b))
>  #define ULONG_CMP_LT(a, b)   (ULONG_MAX / 2 < (a) - (b))
>  
> @@ -425,7 +435,8 @@ static inline void rcu_preempt_sleep_check(void) { }
>  
>  // See RCU_LOCKDEP_WARN() for an explanation of the double call to
>  // debug_lockdep_rcu_enabled().
> -static inline bool lockdep_assert_rcu_helper(bool c)
> +static inline bool lockdep_assert_rcu_helper(bool c, const struct 
> __ctx_guard_RCU *ctx)
> +     __assumes_shared_ctx_guard(RCU) __assumes_shared_ctx_guard(ctx)
>  {
>       return debug_lockdep_rcu_enabled() &&
>              (c || !rcu_is_watching() || !rcu_lockdep_current_cpu_online()) &&
> @@ -438,7 +449,7 @@ static inline bool lockdep_assert_rcu_helper(bool c)
>   * Splats if lockdep is enabled and there is no rcu_read_lock() in effect.
>   */
>  #define lockdep_assert_in_rcu_read_lock() \
> -     WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_lock_map)))
> +     WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_lock_map), 
> RCU))
>  
>  /**
>   * lockdep_assert_in_rcu_read_lock_bh - WARN if not protected by 
> rcu_read_lock_bh()
> @@ -448,7 +459,7 @@ static inline bool lockdep_assert_rcu_helper(bool c)
>   * actual rcu_read_lock_bh() is required.
>   */
>  #define lockdep_assert_in_rcu_read_lock_bh() \
> -     WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_bh_lock_map)))
> +     WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_bh_lock_map), 
> RCU_BH))
>  
>  /**
>   * lockdep_assert_in_rcu_read_lock_sched - WARN if not protected by 
> rcu_read_lock_sched()
> @@ -458,7 +469,7 @@ static inline bool lockdep_assert_rcu_helper(bool c)
>   * instead an actual rcu_read_lock_sched() is required.
>   */
>  #define lockdep_assert_in_rcu_read_lock_sched() \
> -     
> WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_sched_lock_map)))
> +     
> WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_sched_lock_map), 
> RCU_SCHED))
>  
>  /**
>   * lockdep_assert_in_rcu_reader - WARN if not within some type of RCU reader
> @@ -476,17 +487,17 @@ static inline bool lockdep_assert_rcu_helper(bool c)
>       WARN_ON_ONCE(lockdep_assert_rcu_helper(!lock_is_held(&rcu_lock_map) &&  
>                 \
>                                              !lock_is_held(&rcu_bh_lock_map) 
> &&               \
>                                              
> !lock_is_held(&rcu_sched_lock_map) &&            \
> -                                            preemptible()))
> +                                            preemptible(), RCU))
>  
>  #else /* #ifdef CONFIG_PROVE_RCU */
>  
>  #define RCU_LOCKDEP_WARN(c, s) do { } while (0 && (c))
>  #define rcu_sleep_check() do { } while (0)
>  
> -#define lockdep_assert_in_rcu_read_lock() do { } while (0)
> -#define lockdep_assert_in_rcu_read_lock_bh() do { } while (0)
> -#define lockdep_assert_in_rcu_read_lock_sched() do { } while (0)
> -#define lockdep_assert_in_rcu_reader() do { } while (0)
> +#define lockdep_assert_in_rcu_read_lock() __assume_shared_ctx_guard(RCU)
> +#define lockdep_assert_in_rcu_read_lock_bh() 
> __assume_shared_ctx_guard(RCU_BH)
> +#define lockdep_assert_in_rcu_read_lock_sched() 
> __assume_shared_ctx_guard(RCU_SCHED)
> +#define lockdep_assert_in_rcu_reader() __assume_shared_ctx_guard(RCU)
>  
>  #endif /* #else #ifdef CONFIG_PROVE_RCU */
>  
> @@ -506,11 +517,11 @@ static inline bool lockdep_assert_rcu_helper(bool c)
>  #endif /* #else #ifdef __CHECKER__ */
>  
>  #define __unrcu_pointer(p, local)                                    \
> -({                                                                   \
> +context_unsafe(                                                              
> \
>       typeof(*p) *local = (typeof(*p) *__force)(p);                   \
>       rcu_check_sparse(p, __rcu);                                     \
> -     ((typeof(*p) __force __kernel *)(local));                       \
> -})
> +     ((typeof(*p) __force __kernel *)(local))                        \
> +)
>  /**
>   * unrcu_pointer - mark a pointer as not being RCU protected
>   * @p: pointer needing to lose its __rcu property
> @@ -586,7 +597,7 @@ static inline bool lockdep_assert_rcu_helper(bool c)
>   * other macros that it invokes.
>   */
>  #define rcu_assign_pointer(p, v)                                           \
> -do {                                                                       \
> +context_unsafe(                                                            \
>       uintptr_t _r_a_p__v = (uintptr_t)(v);                                 \
>       rcu_check_sparse(p, __rcu);                                           \
>                                                                             \
> @@ -594,7 +605,7 @@ do {                                                      
>                       \
>               WRITE_ONCE((p), (typeof(p))(_r_a_p__v));                      \
>       else                                                                  \
>               smp_store_release(&p, RCU_INITIALIZER((typeof(p))_r_a_p__v)); \
> -} while (0)
> +)
>  
>  /**
>   * rcu_replace_pointer() - replace an RCU pointer, returning its old value
> @@ -861,9 +872,10 @@ do {                                                     
>                       \
>   * only when acquiring spinlocks that are subject to priority inheritance.
>   */
>  static __always_inline void rcu_read_lock(void)
> +     __acquires_shared(RCU)
>  {
>       __rcu_read_lock();
> -     __acquire(RCU);
> +     __acquire_shared(RCU);
>       rcu_lock_acquire(&rcu_lock_map);
>       RCU_LOCKDEP_WARN(!rcu_is_watching(),
>                        "rcu_read_lock() used illegally while idle");
> @@ -891,11 +903,12 @@ static __always_inline void rcu_read_lock(void)
>   * See rcu_read_lock() for more information.
>   */
>  static inline void rcu_read_unlock(void)
> +     __releases_shared(RCU)
>  {
>       RCU_LOCKDEP_WARN(!rcu_is_watching(),
>                        "rcu_read_unlock() used illegally while idle");
>       rcu_lock_release(&rcu_lock_map); /* Keep acq info for rls diags. */
> -     __release(RCU);
> +     __release_shared(RCU);
>       __rcu_read_unlock();
>  }
>  
> @@ -914,9 +927,11 @@ static inline void rcu_read_unlock(void)
>   * was invoked from some other task.
>   */
>  static inline void rcu_read_lock_bh(void)
> +     __acquires_shared(RCU) __acquires_shared(RCU_BH)
>  {
>       local_bh_disable();
> -     __acquire(RCU_BH);
> +     __acquire_shared(RCU);
> +     __acquire_shared(RCU_BH);
>       rcu_lock_acquire(&rcu_bh_lock_map);
>       RCU_LOCKDEP_WARN(!rcu_is_watching(),
>                        "rcu_read_lock_bh() used illegally while idle");
> @@ -928,11 +943,13 @@ static inline void rcu_read_lock_bh(void)
>   * See rcu_read_lock_bh() for more information.
>   */
>  static inline void rcu_read_unlock_bh(void)
> +     __releases_shared(RCU) __releases_shared(RCU_BH)
>  {
>       RCU_LOCKDEP_WARN(!rcu_is_watching(),
>                        "rcu_read_unlock_bh() used illegally while idle");
>       rcu_lock_release(&rcu_bh_lock_map);
> -     __release(RCU_BH);
> +     __release_shared(RCU_BH);
> +     __release_shared(RCU);
>       local_bh_enable();
>  }
>  
> @@ -952,9 +969,11 @@ static inline void rcu_read_unlock_bh(void)
>   * rcu_read_lock_sched() was invoked from an NMI handler.
>   */
>  static inline void rcu_read_lock_sched(void)
> +     __acquires_shared(RCU) __acquires_shared(RCU_SCHED)
>  {
>       preempt_disable();
> -     __acquire(RCU_SCHED);
> +     __acquire_shared(RCU);
> +     __acquire_shared(RCU_SCHED);
>       rcu_lock_acquire(&rcu_sched_lock_map);
>       RCU_LOCKDEP_WARN(!rcu_is_watching(),
>                        "rcu_read_lock_sched() used illegally while idle");
> @@ -962,9 +981,11 @@ static inline void rcu_read_lock_sched(void)
>  
>  /* Used by lockdep and tracing: cannot be traced, cannot call lockdep. */
>  static inline notrace void rcu_read_lock_sched_notrace(void)
> +     __acquires_shared(RCU) __acquires_shared(RCU_SCHED)
>  {
>       preempt_disable_notrace();
> -     __acquire(RCU_SCHED);
> +     __acquire_shared(RCU);
> +     __acquire_shared(RCU_SCHED);
>  }
>  
>  /**
> @@ -973,22 +994,27 @@ static inline notrace void 
> rcu_read_lock_sched_notrace(void)
>   * See rcu_read_lock_sched() for more information.
>   */
>  static inline void rcu_read_unlock_sched(void)
> +     __releases_shared(RCU) __releases_shared(RCU_SCHED)
>  {
>       RCU_LOCKDEP_WARN(!rcu_is_watching(),
>                        "rcu_read_unlock_sched() used illegally while idle");
>       rcu_lock_release(&rcu_sched_lock_map);
> -     __release(RCU_SCHED);
> +     __release_shared(RCU_SCHED);
> +     __release_shared(RCU);
>       preempt_enable();
>  }
>  
>  /* Used by lockdep and tracing: cannot be traced, cannot call lockdep. */
>  static inline notrace void rcu_read_unlock_sched_notrace(void)
> +     __releases_shared(RCU) __releases_shared(RCU_SCHED)
>  {
> -     __release(RCU_SCHED);
> +     __release_shared(RCU_SCHED);
> +     __release_shared(RCU);
>       preempt_enable_notrace();
>  }
>  
>  static __always_inline void rcu_read_lock_dont_migrate(void)
> +     __acquires_shared(RCU)
>  {
>       if (IS_ENABLED(CONFIG_PREEMPT_RCU))
>               migrate_disable();
> @@ -996,6 +1022,7 @@ static __always_inline void 
> rcu_read_lock_dont_migrate(void)
>  }
>  
>  static inline void rcu_read_unlock_migrate(void)
> +     __releases_shared(RCU)
>  {
>       rcu_read_unlock();
>       if (IS_ENABLED(CONFIG_PREEMPT_RCU))
> @@ -1041,10 +1068,10 @@ static inline void rcu_read_unlock_migrate(void)
>   * ordering guarantees for either the CPU or the compiler.
>   */
>  #define RCU_INIT_POINTER(p, v) \
> -     do { \
> +     context_unsafe( \
>               rcu_check_sparse(p, __rcu); \
>               WRITE_ONCE(p, RCU_INITIALIZER(v)); \
> -     } while (0)
> +     )
>  
>  /**
>   * RCU_POINTER_INITIALIZER() - statically initialize an RCU protected pointer
> @@ -1206,4 +1233,6 @@ DEFINE_LOCK_GUARD_0(rcu,
>       } while (0),
>       rcu_read_unlock())
>  
> +DECLARE_LOCK_GUARD_0_ATTRS(rcu, __acquires_shared(RCU), 
> __releases_shared(RCU))
> +
>  #endif /* __LINUX_RCUPDATE_H */
> diff --git a/lib/test_context-analysis.c b/lib/test_context-analysis.c
> index 77e599a9281b..f18b7252646d 100644
> --- a/lib/test_context-analysis.c
> +++ b/lib/test_context-analysis.c
> @@ -7,6 +7,7 @@
>  #include <linux/bit_spinlock.h>
>  #include <linux/build_bug.h>
>  #include <linux/mutex.h>
> +#include <linux/rcupdate.h>
>  #include <linux/seqlock.h>
>  #include <linux/spinlock.h>
>  
> @@ -277,3 +278,87 @@ static void __used test_bit_spin_lock(struct 
> test_bit_spinlock_data *d)
>               bit_spin_unlock(3, &d->bits);
>       }
>  }
> +
> +/*
> + * Test that we can mark a variable guarded by RCU, and we can dereference 
> and
> + * write to the pointer with RCU's primitives.
> + */
> +struct test_rcu_data {
> +     long __rcu_guarded *data;
> +};
> +
> +static void __used test_rcu_guarded_reader(struct test_rcu_data *d)
> +{
> +     rcu_read_lock();
> +     (void)rcu_dereference(d->data);
> +     rcu_read_unlock();
> +
> +     rcu_read_lock_bh();
> +     (void)rcu_dereference(d->data);
> +     rcu_read_unlock_bh();
> +
> +     rcu_read_lock_sched();
> +     (void)rcu_dereference(d->data);
> +     rcu_read_unlock_sched();
> +}
> +
> +static void __used test_rcu_guard(struct test_rcu_data *d)
> +{
> +     guard(rcu)();
> +     (void)rcu_dereference(d->data);
> +}
> +
> +static void __used test_rcu_guarded_updater(struct test_rcu_data *d)
> +{
> +     rcu_assign_pointer(d->data, NULL);
> +     RCU_INIT_POINTER(d->data, NULL);
> +     (void)unrcu_pointer(d->data);
> +}
> +
> +static void wants_rcu_held(void)     __must_hold_shared(RCU)       { }
> +static void wants_rcu_held_bh(void)  __must_hold_shared(RCU_BH)    { }
> +static void wants_rcu_held_sched(void)       __must_hold_shared(RCU_SCHED) { 
> }
> +
> +static void __used test_rcu_lock_variants(void)
> +{
> +     rcu_read_lock();
> +     wants_rcu_held();
> +     rcu_read_unlock();
> +
> +     rcu_read_lock_bh();
> +     wants_rcu_held_bh();
> +     rcu_read_unlock_bh();
> +
> +     rcu_read_lock_sched();
> +     wants_rcu_held_sched();
> +     rcu_read_unlock_sched();
> +}
> +
> +static void __used test_rcu_lock_reentrant(void)
> +{
> +     rcu_read_lock();
> +     rcu_read_lock();
> +     rcu_read_lock_bh();
> +     rcu_read_lock_bh();
> +     rcu_read_lock_sched();
> +     rcu_read_lock_sched();
> +
> +     rcu_read_unlock_sched();
> +     rcu_read_unlock_sched();
> +     rcu_read_unlock_bh();
> +     rcu_read_unlock_bh();
> +     rcu_read_unlock();
> +     rcu_read_unlock();
> +}
> +
> +static void __used test_rcu_assert_variants(void)
> +{
> +     lockdep_assert_in_rcu_read_lock();
> +     wants_rcu_held();
> +
> +     lockdep_assert_in_rcu_read_lock_bh();
> +     wants_rcu_held_bh();
> +
> +     lockdep_assert_in_rcu_read_lock_sched();
> +     wants_rcu_held_sched();
> +}
> -- 
> 2.52.0.rc1.455.g30608eb744-goog
> 

Reply via email to