From: Stephen Hemminger <step...@networkplumber.org> Date: Fri, 1 Dec 2017 15:30:01 -0800
> On Fri, 1 Dec 2017 12:11:56 -0800 > Stephen Hemminger <step...@networkplumber.org> wrote: > >> This is another way of addressing the GSO maximum performance issues for >> containers on Azure. What happens is that the underlying infrastructure uses >> a overlay network such that GSO packets over 64K - vlan header end up cause >> either guest or host to have do expensive software copy and fragmentation. >> >> The netvsc driver reports GSO maximum settings correctly, the issue >> is that containers on veth devices still have the larger settings. >> One solution that was examined was propogating the values back >> through the bridge device, but this does not work for cases where >> virtual container network is done on L3. >> >> This patch set punts the problem to the orchestration layer that sets >> up the container network. It also enables other virtual devices >> to have configurable settings for GSO maximum. >> >> Stephen Hemminger (2): >> rtnetlink: allow GSO maximums to be passed to device >> veth: allow configuring GSO maximums >> >> drivers/net/veth.c | 20 ++++++++++++++++++++ >> net/core/rtnetlink.c | 2 ++ >> 2 files changed, 22 insertions(+) >> > > I would like a confirmation from Intel that is doing Docker testing > that this works for them before merging. Like David Ahern, I think you should allow this net netlink setting during changelink as well as newlink. Thanks.