On 8/22/17 6:40 PM, Alexei Starovoitov wrote: >> diff --git a/kernel/cgroup/cgroup.c b/kernel/cgroup/cgroup.c >> index df2e0f14a95d..7480cebab073 100644 >> --- a/kernel/cgroup/cgroup.c >> +++ b/kernel/cgroup/cgroup.c >> @@ -5186,4 +5186,22 @@ int cgroup_bpf_update(struct cgroup *cgrp, struct >> bpf_prog *prog, >> mutex_unlock(&cgroup_mutex); >> return ret; >> } >> + >> +int cgroup_bpf_run_filter_sk(struct sock *sk, >> + enum bpf_attach_type type) >> +{ >> + struct cgroup *cgrp = sock_cgroup_ptr(&sk->sk_cgrp_data); >> + int ret = 0; >> + >> + while (cgrp) { >> + ret = __cgroup_bpf_run_filter_sk(cgrp, sk, type); >> + if (ret < 0) >> + break; >> + >> + cgrp = cgroup_parent(cgrp); >> + } > > I think this walk changes semantics for existing setups, so we cannot do it > by default and have to add new attach flag.
I can add a flag similar to the override. > Also why break on (ret < 0) ? Because __cgroup_bpf_run_filter_sk returns either 0 or -EPERM. > The caller of this does: > err = BPF_CGROUP_RUN_PROG_INET_SOCK(sk); > if (err) { > sk_common_release(sk); > so we should probably break out of the loop on if (ret) too. > I'll do that in v2.