Make kprobes to accept 1-level nesting instead of missing it
on arm64.

Any kprobes hits in kprobes pre/post handler context can be
nested at once. If the other kprobes hits in the nested pre/post
handler context, that will be missed.

We can test this feature on the kernel with
CONFIG_KPROBE_EVENTS_ON_NOTRACE=y as below.

 # cd /sys/kernel/debug/tracing
 # echo p ring_buffer_lock_reserve > kprobe_events
 # echo p vfs_read >> kprobe_events
 # echo stacktrace > events/kprobes/p_ring_buffer_lock_reserve_0/trigger
 # echo 1 > events/kprobes/enable
 # cat trace
 ...
               sh-211   [005] d..2    71.708242: p_ring_buffer_lock_reserve_0: 
(ring_buffer_lock_reserve+0x0/0x4c8)
               sh-211   [005] d..2    71.709982: <stack trace>
  => kprobe_dispatcher
  => kprobe_breakpoint_handler
  => call_break_hook
  => brk_handler
  => do_debug_exception
  => el1_sync_handler
  => el1_sync
  => ring_buffer_lock_reserve
  => kprobe_trace_func
  => kprobe_dispatcher
  => kprobe_breakpoint_handler
  => call_break_hook
  => brk_handler
  => do_debug_exception
  => el1_sync_handler
  => el1_sync
  => vfs_read
  => __arm64_sys_read
  => el0_svc_common.constprop.3
  => do_el0_svc
  => el0_sync_handler
  => el0_sync

This shows brk_handler is nested.

Note that this also improve unrecoverable message to show
nested probes too.

Signed-off-by: Masami Hiramatsu <mhira...@kernel.org>
---
  Changes in v2:
   - Dump nested kprobes when hit a BUG().
---
 arch/arm64/include/asm/kprobes.h   |    5 ++
 arch/arm64/kernel/probes/kprobes.c |   79 +++++++++++++++++++++---------------
 2 files changed, 50 insertions(+), 34 deletions(-)

diff --git a/arch/arm64/include/asm/kprobes.h b/arch/arm64/include/asm/kprobes.h
index 97e511d645a2..b2ebd3bad794 100644
--- a/arch/arm64/include/asm/kprobes.h
+++ b/arch/arm64/include/asm/kprobes.h
@@ -34,12 +34,15 @@ struct kprobe_step_ctx {
        unsigned long match_addr;
 };
 
+#define KPROBE_NEST_MAX 2
+
 /* per-cpu kprobe control block */
 struct kprobe_ctlblk {
        unsigned int kprobe_status;
        unsigned long saved_irqflag;
-       struct prev_kprobe prev_kprobe;
+       struct prev_kprobe prev[KPROBE_NEST_MAX];
        struct kprobe_step_ctx ss_ctx;
+       int nested;
 };
 
 void arch_remove_kprobe(struct kprobe *);
diff --git a/arch/arm64/kernel/probes/kprobes.c 
b/arch/arm64/kernel/probes/kprobes.c
index d1c95dcf1d78..1da3df07bcd4 100644
--- a/arch/arm64/kernel/probes/kprobes.c
+++ b/arch/arm64/kernel/probes/kprobes.c
@@ -153,14 +153,18 @@ void __kprobes arch_remove_kprobe(struct kprobe *p)
 
 static void __kprobes save_previous_kprobe(struct kprobe_ctlblk *kcb)
 {
-       kcb->prev_kprobe.kp = kprobe_running();
-       kcb->prev_kprobe.status = kcb->kprobe_status;
+       int i = kcb->nested++;
+
+       kcb->prev[i].kp = kprobe_running();
+       kcb->prev[i].status = kcb->kprobe_status;
 }
 
 static void __kprobes restore_previous_kprobe(struct kprobe_ctlblk *kcb)
 {
-       __this_cpu_write(current_kprobe, kcb->prev_kprobe.kp);
-       kcb->kprobe_status = kcb->prev_kprobe.status;
+       int i = --kcb->nested;
+
+       __this_cpu_write(current_kprobe, kcb->prev[i].kp);
+       kcb->kprobe_status = kcb->prev[i].status;
 }
 
 static void __kprobes set_current_kprobe(struct kprobe *p)
@@ -168,6 +172,14 @@ static void __kprobes set_current_kprobe(struct kprobe *p)
        __this_cpu_write(current_kprobe, p);
 }
 
+static nokprobe_inline void pop_current_kprobe(struct kprobe_ctlblk *kcb)
+{
+       if (kcb->nested)
+               restore_previous_kprobe(kcb);
+       else
+               reset_current_kprobe();
+}
+
 /*
  * Interrupts need to be disabled before single-step mode is set, and not
  * reenabled until after single-step mode ends.
@@ -243,13 +255,21 @@ static int __kprobes reenter_kprobe(struct kprobe *p,
        switch (kcb->kprobe_status) {
        case KPROBE_HIT_SSDONE:
        case KPROBE_HIT_ACTIVE:
+               if (kcb->nested < KPROBE_NEST_MAX - 1) {
+                       save_previous_kprobe(kcb);
+                       return 0;
+               }
                kprobes_inc_nmissed_count(p);
                setup_singlestep(p, regs, kcb, 1);
                break;
        case KPROBE_HIT_SS:
        case KPROBE_REENTER:
                pr_warn("Unrecoverable kprobe detected.\n");
+               pr_err("Current kprobe:\n");
                dump_kprobe(p);
+               pr_err("Nested kprobes (nested: %d):\n", kcb->nested);
+               while (kcb->nested)
+                       dump_kprobe(kcb->prev[--kcb->nested].kp);
                BUG();
                break;
        default:
@@ -286,7 +306,7 @@ post_kprobe_handler(struct kprobe_ctlblk *kcb, struct 
pt_regs *regs)
                cur->post_handler(cur, regs, 0);
        }
 
-       reset_current_kprobe();
+       pop_current_kprobe(kcb);
 }
 
 int __kprobes kprobe_fault_handler(struct pt_regs *regs, unsigned int fsr)
@@ -309,11 +329,7 @@ int __kprobes kprobe_fault_handler(struct pt_regs *regs, 
unsigned int fsr)
                        BUG();
 
                kernel_disable_single_step();
-
-               if (kcb->kprobe_status == KPROBE_REENTER)
-                       restore_previous_kprobe(kcb);
-               else
-                       reset_current_kprobe();
+               pop_current_kprobe(kcb);
 
                break;
        case KPROBE_HIT_ACTIVE:
@@ -357,30 +373,27 @@ static void __kprobes kprobe_handler(struct pt_regs *regs)
        p = get_kprobe((kprobe_opcode_t *) addr);
 
        if (p) {
-               if (cur_kprobe) {
-                       if (reenter_kprobe(p, regs, kcb))
-                               return;
-               } else {
-                       /* Probe hit */
-                       set_current_kprobe(p);
-                       kcb->kprobe_status = KPROBE_HIT_ACTIVE;
+               if (cur_kprobe && reenter_kprobe(p, regs, kcb))
+                       return;
+               /* Probe hit */
+               set_current_kprobe(p);
+               kcb->kprobe_status = KPROBE_HIT_ACTIVE;
 
-                       /*
-                        * If we have no pre-handler or it returned 0, we
-                        * continue with normal processing.  If we have a
-                        * pre-handler and it returned non-zero, it will
-                        * modify the execution path and no need to single
-                        * stepping. Let's just reset current kprobe and exit.
-                        *
-                        * pre_handler can hit a breakpoint and can step thru
-                        * before return, keep PSTATE D-flag enabled until
-                        * pre_handler return back.
-                        */
-                       if (!p->pre_handler || !p->pre_handler(p, regs)) {
-                               setup_singlestep(p, regs, kcb, 0);
-                       } else
-                               reset_current_kprobe();
-               }
+               /*
+                * If we have no pre-handler or it returned 0, we
+                * continue with normal processing.  If we have a
+                * pre-handler and it returned non-zero, it will
+                * modify the execution path and no need to single
+                * stepping. Let's just reset current kprobe and exit.
+                *
+                * pre_handler can hit a breakpoint and can step thru
+                * before return, keep PSTATE D-flag enabled until
+                * pre_handler return back.
+                */
+               if (!p->pre_handler || !p->pre_handler(p, regs)) {
+                       setup_singlestep(p, regs, kcb, 0);
+               } else
+                       pop_current_kprobe(kcb);
        }
        /*
         * The breakpoint instruction was removed right

Reply via email to