Hi Catalin,

Thanks for your review :D

> On Fri, Feb 27, 2026 at 03:17:02PM +0000, Yeoreum Yun wrote:
> > +#ifdef CONFIG_ARM64_LSUI
> > +
> > +/*
> > + * FEAT_LSUI is supported since Armv9.6, where FEAT_PAN is mandatory.
> > + * However, this assumption may not always hold:
> > + *
> > + *   - Some CPUs advertise FEAT_LSUI but lack FEAT_PAN.
> > + *   - Virtualisation or ID register overrides may expose invalid
> > + *     feature combinations.
> > + *
> > + * Rather than disabling FEAT_LSUI when FEAT_PAN is absent, wrap LSUI
> > + * instructions with uaccess_ttbr0_enable()/disable() when
> > + * ARM64_SW_TTBR0_PAN is enabled.
> > + */
>
> I'd just keep this comment in the commit log. Here you could simply say
> that user access instructions don't require (hardware) PAN toggling. It
> should be obvious why we use ttbr0 toggling like for other uaccess
> routines.

Okay. I'll move this into commit log. Thanks!

>
> > +#define LSUI_FUTEX_ATOMIC_OP(op, asm_op)                           \
> > +static __always_inline int                                         \
> > +__lsui_futex_atomic_##op(int oparg, u32 __user *uaddr, int *oval)  \
> > +{                                                                  \
> > +   int ret = 0;                                                    \
> > +   int oldval;                                                     \
> > +                                                                   \
> > +   uaccess_ttbr0_enable();                                         \
> > +                                                                   \
> > +   asm volatile("// __lsui_futex_atomic_" #op "\n"                 \
> > +   __LSUI_PREAMBLE                                                 \
> > +"1:        " #asm_op "al   %w3, %w2, %1\n"                                 
> > \
>
> As I mentioned on a previous patch, can we not use named operators here?

I missed your message before I sent to v16, But v16 already make them
with named operands. Thanks!

[...]

> > +}
> > +
> > +static __always_inline int
> > +__lsui_cmpxchg32(u32 __user *uaddr, u32 oldval, u32 newval, u32 *oval)
> > +{
> > +   u64 __user *uaddr64;
> > +   bool futex_pos, other_pos;
> > +   int ret, i;
> > +   u32 other, orig_other;
> > +   union {
> > +           u32 futex[2];
> > +           u64 raw;
> > +   } oval64, orig64, nval64;
> > +
> > +   uaddr64 = (u64 __user *) PTR_ALIGN_DOWN(uaddr, sizeof(u64));
>
> Nit: we don't use space after the type cast.

Oops. I'll remove space.

>
> > +   futex_pos = !IS_ALIGNED((unsigned long)uaddr, sizeof(u64));
> > +   other_pos = !futex_pos;
> > +
> > +   oval64.futex[futex_pos] = oldval;
> > +   ret = get_user(oval64.futex[other_pos], (u32 __user *)uaddr64 + 
> > other_pos);
> > +   if (ret)
> > +           return -EFAULT;
> > +
> > +   ret = -EAGAIN;
> > +   for (i = 0; i < FUTEX_MAX_LOOPS; i++) {
>
> I was wondering if we still need the FUTEX_MAX_LOOPS bound with LSUI. I
> guess with CAS we can have some malicious user that keeps updating the
> futex location or the adjacent one on another CPU. However, I think we'd
> need to differentiate between futex_atomic_cmpxchg_inatomic() use and
> the eor case.

Hmm. I'll comment below together in eor..

>
> > +           orig64.raw = nval64.raw = oval64.raw;
> > +
> > +           nval64.futex[futex_pos] = newval;
>
> I'd keep orig64.raw = oval64.raw and set the nval64 separately (I find
> it clearer, not sure the compiler cares much):
>
>               nval64.futex[futex_pos] = newval;
>               nval64.futex[other_pos] = oval64.futex[other_pos];
>
> > +
> > +           if (__lsui_cmpxchg64(uaddr64, &oval64.raw, nval64.raw))
> > +                   return -EFAULT;
> > +
> > +           oldval = oval64.futex[futex_pos];
> > +           other = oval64.futex[other_pos];
> > +           orig_other = orig64.futex[other_pos];
> > +
> > +           if (other == orig_other) {
> > +                   ret = 0;
> > +                   break;
> > +           }
>
> Is this check correct? What if the cmpxchg64 failed because futex_pos
> was changed but other_pos remained the same, it will just report success
> here. You need to compare the full 64-bit value to ensure the cmpxchg64
> succeeded.

This is not matter since "futex_cmpxchg_value_locked()" checks
the "curval" and "oldval" IOW, though it returns success,
caller of this function always checks the "curval" and "oldval"
and when it's different, It handles to change return as -EAGAIN.

>
> > +   }
> > +
> > +   if (!ret)
> > +           *oval = oldval;
> > +
> > +   return ret;
> > +}
> > +
> > +static __always_inline int
> > +__lsui_futex_atomic_and(int oparg, u32 __user *uaddr, int *oval)
> > +{
> > +   /*
> > +    * Undo the bitwise negation applied to the oparg passed from
> > +    * arch_futex_atomic_op_inuser() with FUTEX_OP_ANDN.
> > +    */
> > +   return __lsui_futex_atomic_andnot(~oparg, uaddr, oval);
> > +}
> > +
> > +static __always_inline int
> > +__lsui_futex_atomic_eor(int oparg, u32 __user *uaddr, int *oval)
> > +{
> > +   u32 oldval, newval, val;
> > +   int ret, i;
> > +
> > +   if (get_user(oldval, uaddr))
> > +           return -EFAULT;
> > +
> > +   /*
> > +    * there are no ldteor/stteor instructions...
> > +    */
> > +   for (i = 0; i < FUTEX_MAX_LOOPS; i++) {
> > +           newval = oldval ^ oparg;
> > +
> > +           ret = __lsui_cmpxchg32(uaddr, oldval, newval, &val);
>
> Since we have a FUTEX_MAX_LOOPS here, do we need it in cmpxchg32 as
> well?
>
> For eor, we need a loop irrespective of whether futex_pos or other_pos
> have changed. For cmpxchg, we need the loop only if other_pos has
> changed and return -EAGAIN if futex_pos has changed since the caller
> needs to update oldval and call again.
>
> So try to differentiate these cases, maybe only keep the loop outside
> cmpxchg32 (I haven't put much though into it).

I think we can remove loops on __lsui_cmpxchg32() and return -EAGAIN
when other_pos is different. the __lsui_cmpxchg32() will be called
"futex_cmpxchg_value_locked()" and as I said, this always checks
whether curval & oldval when it successed.

But in "eor" when it receive "-EAGAIN" from __lsui_cmxchg32()
we can simply continue the loop.

>
> > +           if (ret)
> > +                   return ret;
> > +
> > +           if (val == oldval) {
> > +                   *oval = val;
> > +                   return 0;
> > +           }
>
> I can see you are adding another check here for the actual value which
> solves the other_pos comparison earlier but that's only for eor and not
> the __lsui_futex_cmpxchg() case.

As I mention above, though it success, caller of futex who calls
__lsui_futex_cmpxchg() via "futex_cmpxchg_value_locked()" checks
curval and oldval is the same even on the success.

So it's not a matter.

>
> > +
> > +           oldval = val;
> > +   }
> > +
> > +   return -EAGAIN;
> > +}
> > +
> > +static __always_inline int
> > +__lsui_futex_cmpxchg(u32 __user *uaddr, u32 oldval, u32 newval, u32 *oval)
> > +{
> > +   return __lsui_cmpxchg32(uaddr, oldval, newval, oval);
> > +}
> > +#endif     /* CONFIG_ARM64_LSUI */
>
> --
> Catalin

Thanks!

--
Sincerely,
Yeoreum Yun

Reply via email to