David Ahern <dsah...@kernel.org> writes: > Add BPF_XDP_DEVMAP attach type for use with programs associated with a > DEVMAP. > > DEVMAPs can associate a program with a device entry by setting the > value to <index,id> pair. The program associated with the id must have > type XDP with expected attach type BPF_XDP_DEVMAP. When a program is > associated with a device index, the program is run on an XDP_REDIRECT > and before the buffer is added to the per-cpu queue. At this point > rxq data is still valid; the next patch adds tx device information > allowing the prorgam to see both ingress and egress device indices. > > XDP generic is skb based and XDP programs do not work with skb's. Block > the use case by walking maps used by a program that is to be attached > via xdpgeneric and fail if any of them are DEVMAP / DEVMAP_HASH with > 8-bytes values. > > Signed-off-by: David Ahern <dsah...@kernel.org> > --- > include/linux/bpf.h | 5 +++ > include/uapi/linux/bpf.h | 1 + > kernel/bpf/devmap.c | 82 +++++++++++++++++++++++++++++++--- > net/core/dev.c | 18 ++++++++ > tools/include/uapi/linux/bpf.h | 1 + > 5 files changed, 102 insertions(+), 5 deletions(-) > > diff --git a/include/linux/bpf.h b/include/linux/bpf.h > index efe8836b5c48..088751bc09aa 100644 > --- a/include/linux/bpf.h > +++ b/include/linux/bpf.h > @@ -1242,6 +1242,7 @@ int dev_map_enqueue(struct bpf_dtab_netdev *dst, struct > xdp_buff *xdp, > struct net_device *dev_rx); > int dev_map_generic_redirect(struct bpf_dtab_netdev *dst, struct sk_buff > *skb, > struct bpf_prog *xdp_prog); > +bool dev_map_can_have_prog(struct bpf_map *map); > > struct bpf_cpu_map_entry *__cpu_map_lookup_elem(struct bpf_map *map, u32 > key); > void __cpu_map_flush(void); > @@ -1355,6 +1356,10 @@ static inline struct net_device > *__dev_map_hash_lookup_elem(struct bpf_map *map > { > return NULL; > } > +static inline bool dev_map_can_have_prog(struct bpf_map *map) > +{ > + return false; > +} > > static inline void __dev_flush(void) > { > diff --git a/include/uapi/linux/bpf.h b/include/uapi/linux/bpf.h > index 97e1fd19ff58..8c2c0d0c9a0e 100644 > --- a/include/uapi/linux/bpf.h > +++ b/include/uapi/linux/bpf.h > @@ -224,6 +224,7 @@ enum bpf_attach_type { > BPF_CGROUP_INET6_GETPEERNAME, > BPF_CGROUP_INET4_GETSOCKNAME, > BPF_CGROUP_INET6_GETSOCKNAME, > + BPF_XDP_DEVMAP, > __MAX_BPF_ATTACH_TYPE > }; > > diff --git a/kernel/bpf/devmap.c b/kernel/bpf/devmap.c > index 2c01ce434306..06f4c746fa7c 100644 > --- a/kernel/bpf/devmap.c > +++ b/kernel/bpf/devmap.c > @@ -70,6 +70,7 @@ struct bpf_dtab_netdev { > struct net_device *dev; /* must be first member, due to tracepoint */ > struct hlist_node index_hlist; > struct bpf_dtab *dtab; > + struct bpf_prog *xdp_prog; > struct rcu_head rcu; > unsigned int idx; > struct dev_map_ext_val val; > @@ -228,6 +229,8 @@ static void dev_map_free(struct bpf_map *map) > > hlist_for_each_entry_safe(dev, next, head, index_hlist) > { > hlist_del_rcu(&dev->index_hlist); > + if (dev->xdp_prog) > + bpf_prog_put(dev->xdp_prog); > dev_put(dev->dev); > kfree(dev); > } > @@ -242,6 +245,8 @@ static void dev_map_free(struct bpf_map *map) > if (!dev) > continue; > > + if (dev->xdp_prog) > + bpf_prog_put(dev->xdp_prog); > dev_put(dev->dev); > kfree(dev); > } > @@ -328,6 +333,16 @@ static int dev_map_hash_get_next_key(struct bpf_map > *map, void *key, > return -ENOENT; > } > > +bool dev_map_can_have_prog(struct bpf_map *map) > +{ > + if ((map->map_type == BPF_MAP_TYPE_DEVMAP || > + map->map_type == BPF_MAP_TYPE_DEVMAP_HASH) && > + map->value_size == 8) > + return true; > + > + return false; > +} > + > static int bq_xmit_all(struct xdp_dev_bulk_queue *bq, u32 flags) > { > struct net_device *dev = bq->dev; > @@ -452,6 +467,39 @@ static inline int __xdp_enqueue(struct net_device *dev, > struct xdp_buff *xdp, > return bq_enqueue(dev, xdpf, dev_rx); > } > > +static struct xdp_buff *dev_map_run_prog(struct net_device *dev, > + struct xdp_buff *xdp, > + struct bpf_prog *xdp_prog) > +{ > + u32 act; > + > + act = bpf_prog_run_xdp(xdp_prog, xdp); > + switch (act) { > + case XDP_DROP: > + fallthrough; > + case XDP_PASS: > + break; > + case XDP_TX: > + fallthrough; > + case XDP_REDIRECT: > + fallthrough; > + default: > + bpf_warn_invalid_xdp_action(act); > + fallthrough; > + case XDP_ABORTED: > + trace_xdp_exception(dev, xdp_prog, act); > + act = XDP_DROP; > + break; > + } > + > + if (act == XDP_DROP) { > + xdp_return_buff(xdp); > + xdp = NULL; > + } > + > + return xdp; > +} > + > int dev_xdp_enqueue(struct net_device *dev, struct xdp_buff *xdp, > struct net_device *dev_rx) > { > @@ -463,6 +511,11 @@ int dev_map_enqueue(struct bpf_dtab_netdev *dst, struct > xdp_buff *xdp, > { > struct net_device *dev = dst->dev; > > + if (dst->xdp_prog) { > + xdp = dev_map_run_prog(dev, xdp, dst->xdp_prog); > + if (!xdp) > + return 0; > + } > return __xdp_enqueue(dev, xdp, dev_rx); > } > > @@ -499,6 +552,8 @@ static void __dev_map_entry_free(struct rcu_head *rcu) > struct bpf_dtab_netdev *dev; > > dev = container_of(rcu, struct bpf_dtab_netdev, rcu); > + if (dev->xdp_prog) > + bpf_prog_put(dev->xdp_prog); > dev_put(dev->dev); > kfree(dev); > } > @@ -549,9 +604,10 @@ static int dev_map_hash_delete_elem(struct bpf_map *map, > void *key) > > static struct bpf_dtab_netdev *__dev_map_alloc_node(struct net *net, > struct bpf_dtab *dtab, > - u32 ifindex, > + u32 ifindex, u32 prog_id, > unsigned int idx) > { > + struct bpf_prog *prog = NULL; > struct bpf_dtab_netdev *dev; > > dev = kmalloc_node(sizeof(*dev), GFP_ATOMIC | __GFP_NOWARN, > @@ -563,12 +619,23 @@ static struct bpf_dtab_netdev > *__dev_map_alloc_node(struct net *net, > if (!dev->dev) > goto err_out; > > + if (prog_id) { > + prog = bpf_prog_by_id(prog_id); > + if (IS_ERR(prog) || prog->type != BPF_PROG_TYPE_XDP || > + prog->expected_attach_type != BPF_XDP_DEVMAP)
If the second or third condition fails here you never bpf_prog_put() the prog ref. > + goto err_dev; > + } > + > dev->idx = idx; > dev->dtab = dtab; > > + dev->xdp_prog = prog; > dev->val.ifindex = ifindex; > + dev->val.prog_id = prog_id; > > return dev; > +err_dev: > + dev_put(dev->dev); > err_out: > kfree(dev); > return ERR_PTR(-EINVAL); > @@ -580,7 +647,7 @@ static int __dev_map_update_elem(struct net *net, struct > bpf_map *map, > struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); > struct bpf_dtab_netdev *dev, *old_dev; > u32 i = *(u32 *)key; > - u32 ifindex; > + u32 ifindex, id = 0; > > if (map->value_size == 4) { > ifindex = *(u32 *)value; > @@ -588,6 +655,7 @@ static int __dev_map_update_elem(struct net *net, struct > bpf_map *map, > struct dev_map_ext_val *val = value; > > ifindex = val->ifindex; > + id = val->prog_id; > } > > if (unlikely(map_flags > BPF_EXIST)) > @@ -599,8 +667,11 @@ static int __dev_map_update_elem(struct net *net, struct > bpf_map *map, > > if (!ifindex) { > dev = NULL; > + /* can not specify id if ifindex is 0 */ > + if (id) > + return -EINVAL; > } else { > - dev = __dev_map_alloc_node(net, dtab, ifindex, i); > + dev = __dev_map_alloc_node(net, dtab, ifindex, id, i); > if (IS_ERR(dev)) > return PTR_ERR(dev); > } > @@ -630,8 +701,8 @@ static int __dev_map_hash_update_elem(struct net *net, > struct bpf_map *map, > struct bpf_dtab_netdev *dev, *old_dev; > u32 idx = *(u32 *)key; > unsigned long flags; > + u32 ifindex, id = 0; > int err = -EEXIST; > - u32 ifindex; > > if (map->value_size == 4) { > ifindex = *(u32 *)value; > @@ -639,6 +710,7 @@ static int __dev_map_hash_update_elem(struct net *net, > struct bpf_map *map, > struct dev_map_ext_val *val = value; > > ifindex = val->ifindex; > + id = val->prog_id; > } > > if (unlikely(map_flags > BPF_EXIST || !ifindex)) > @@ -650,7 +722,7 @@ static int __dev_map_hash_update_elem(struct net *net, > struct bpf_map *map, > if (old_dev && (map_flags & BPF_NOEXIST)) > goto out_err; > > - dev = __dev_map_alloc_node(net, dtab, ifindex, idx); > + dev = __dev_map_alloc_node(net, dtab, ifindex, id, idx); > if (IS_ERR(dev)) { > err = PTR_ERR(dev); > goto out_err; > diff --git a/net/core/dev.c b/net/core/dev.c > index f36bd3b21997..2571a8976cd4 100644 > --- a/net/core/dev.c > +++ b/net/core/dev.c > @@ -5410,6 +5410,18 @@ static int generic_xdp_install(struct net_device *dev, > struct netdev_bpf *xdp) > struct bpf_prog *new = xdp->prog; > int ret = 0; > > + if (new) { > + u32 i; > + > + /* generic XDP does not work with DEVMAPs that can > + * have a bpf_prog installed > + */ > + for (i = 0; i < new->aux->used_map_cnt; i++) { > + if (dev_map_can_have_prog(new->aux->used_maps[i])) > + return -EINVAL; > + } > + } > + > switch (xdp->command) { > case XDP_SETUP_PROG: > rcu_assign_pointer(dev->xdp_prog, new); > @@ -8825,6 +8837,12 @@ int dev_change_xdp_fd(struct net_device *dev, struct > netlink_ext_ack *extack, > return -EINVAL; > } > > + if (prog->expected_attach_type == BPF_XDP_DEVMAP) { > + NL_SET_ERR_MSG(extack, "BPF_XDP_DEVMAP programs can not > be attached to a device"); > + bpf_prog_put(prog); > + return -EINVAL; > + } > + > /* prog->aux->id may be 0 for orphaned device-bound progs */ > if (prog->aux->id && prog->aux->id == prog_id) { > bpf_prog_put(prog); > diff --git a/tools/include/uapi/linux/bpf.h b/tools/include/uapi/linux/bpf.h > index 97e1fd19ff58..8c2c0d0c9a0e 100644 > --- a/tools/include/uapi/linux/bpf.h > +++ b/tools/include/uapi/linux/bpf.h > @@ -224,6 +224,7 @@ enum bpf_attach_type { > BPF_CGROUP_INET6_GETPEERNAME, > BPF_CGROUP_INET4_GETSOCKNAME, > BPF_CGROUP_INET6_GETSOCKNAME, > + BPF_XDP_DEVMAP, > __MAX_BPF_ATTACH_TYPE > }; > > -- > 2.21.1 (Apple Git-122.3)