On 02/19/2019 06:38 AM, brakmo wrote:
> This patch adds a new bpf helper BPF_FUNC_tcp_check_probe_timer
> "int bpf_check_tcp_probe_timer(struct tcp_bpf_sock *tp, u32 when_us)".
> It is added to BPF_PROG_TYPE_CGROUP_SKB typed bpf_prog which currently
> can be attached to the ingress and egress path.
> 
> The function forces when_us to be at least TCP_TIMEOUT_MIN (currently
> 2 jiffies) and no more than TCP_RTO_MIN (currently 200ms).
> 
> When using a bpf_prog to limit the egress bandwidth of a cgroup,
> it can happen that we drop a packet of a connection that has no
> packets out. In this case, the connection may not retry sending
> the packet until the probe timer fires. Since the default value
> of the probe timer is at least 200ms, this can introduce link
> underutiliation (i.e. the cgroup egress bandwidth being smaller
> than the specified rate) thus increased tail latency.
> This helper function allows for setting a smaller probe timer.
> 
> Signed-off-by: Lawrence Brakmo <bra...@fb.com>
> ---
>  include/uapi/linux/bpf.h | 12 +++++++++++-
>  net/core/filter.c        | 27 +++++++++++++++++++++++++++
>  2 files changed, 38 insertions(+), 1 deletion(-)
> 
> diff --git a/include/uapi/linux/bpf.h b/include/uapi/linux/bpf.h
> index 5daf404511f7..a78936acccae 100644
> --- a/include/uapi/linux/bpf.h
> +++ b/include/uapi/linux/bpf.h
> @@ -2372,6 +2372,15 @@ union bpf_attr {
>   *           val should be one of 0, 1, 2, 3.
>   *   Return
>   *           -EINVAL on error (e.g. val > 3), 0 otherwise.
> + *
> + * int bpf_tcp_check_probe_timer(struct bpf_tcp_sock *tp, int when_us)
> + *   Description
> + *           Checks that there are no packets out and there is no pending
> + *           timer. If both of these are true, it bounds when_us by
> + *           TCP_TIMEOUT_MIN (2 jiffies) or TCP_RTO_MIN (200ms) and
> + *           sets the probe timer.
> + *   Return
> + *           0
>   */
>  #define __BPF_FUNC_MAPPER(FN)                \
>       FN(unspec),                     \
> @@ -2472,7 +2481,8 @@ union bpf_attr {
>       FN(sk_fullsock),                \
>       FN(tcp_sock),                   \
>       FN(tcp_enter_cwr),              \
> -     FN(skb_set_ecn),
> +     FN(skb_set_ecn),                \
> +     FN(tcp_check_probe_timer),
>  
>  /* integer value in 'imm' field of BPF_CALL instruction selects which helper
>   * function eBPF program intends to call
> diff --git a/net/core/filter.c b/net/core/filter.c
> index 275acfb2117d..2b975e651a04 100644
> --- a/net/core/filter.c
> +++ b/net/core/filter.c
> @@ -5465,6 +5465,31 @@ static const struct bpf_func_proto 
> bpf_skb_set_ecn_proto = {
>       .arg1_type      = ARG_PTR_TO_CTX,
>       .arg2_type      = ARG_ANYTHING,
>  };
> +
> +BPF_CALL_2(bpf_tcp_check_probe_timer, struct tcp_sock *, tp, u32, when_us)
> +{
> +     struct sock *sk = (struct sock *) tp;
> +     unsigned long when = usecs_to_jiffies(when_us);
> +
> +     if (!tp->packets_out && !inet_csk(sk)->icsk_pending) {
> +             if (when < TCP_TIMEOUT_MIN)
> +                     when = TCP_TIMEOUT_MIN;
> +             else if (when > TCP_RTO_MIN)
> +                     when = TCP_RTO_MIN;
> +
> +             inet_csk_reset_xmit_timer(sk, ICSK_TIME_PROBE0,
> +                                       when, TCP_RTO_MAX);

Should this be using tcp_reset_xmit_timer() instead to take pacing
into account? (If not, would be good to have a comment explaining
why it's okay to use directly here.)

> +     }
> +     return 0;
> +}
> +
> +static const struct bpf_func_proto bpf_tcp_check_probe_timer_proto = {
> +     .func           = bpf_tcp_check_probe_timer,
> +     .gpl_only       = false,
> +     .ret_type       = RET_INTEGER,
> +     .arg1_type      = ARG_PTR_TO_TCP_SOCK,
> +     .arg2_type      = ARG_ANYTHING,
> +};
>  #endif /* CONFIG_INET */
>  
>  bool bpf_helper_changes_pkt_data(void *func)
> @@ -5628,6 +5653,8 @@ cg_skb_func_proto(enum bpf_func_id func_id, const 
> struct bpf_prog *prog)
>               return &bpf_tcp_enter_cwr_proto;
>       case BPF_FUNC_skb_set_ecn:
>               return &bpf_skb_set_ecn_proto;
> +     case BPF_FUNC_tcp_check_probe_timer:
> +             return &bpf_tcp_check_probe_timer_proto;
>  #endif
>       default:
>               return sk_filter_func_proto(func_id, prog);
> 

Reply via email to