David Miller
2017-May-09 17:48 UTC
[Bridge] [PATCH 0/2] net: Set maximum receive packet size on veth interfaces
From: Fredrik Markstrom <fredrik.markstrom at gmail.com> Date: Tue, 9 May 2017 14:44:36 +0200> Currently veth drops all packets larger then the mtu set on the > receiving end of the pair. This is inconsistent with most hardware > ethernet drivers.False. In fact, many pieces of ethernet hardware a not physically capable of sending even VLAN packets that are above the normal base ethernet MTU. It is therefore untenable to remove this restriction univerally like this.
Fredrik Markström
2017-May-10 00:38 UTC
[Bridge] [PATCH 0/2] net: Set maximum receive packet size on veth interfaces
On Tue, May 9, 2017 at 7:48 PM, David Miller <davem at davemloft.net> wrote:> > From: Fredrik Markstrom <fredrik.markstrom at gmail.com> > Date: Tue, 9 May 2017 14:44:36 +0200 > > > Currently veth drops all packets larger then the mtu set on the > > receiving end of the pair. This is inconsistent with most hardware > > ethernet drivers. > > False. > > In fact, many pieces of ethernet hardware a not physically capable of > sending even VLAN packets that are above the normal base ethernet MTU. >Maybe I was unclear, the veth implementation drops all packers larger then the configured MTU (on the receiving interface). Most ethernet drivers accepts packets up to the ethernet MTU no matter the configured MTU. As far as I can tell from the RFC:s that is ok. A simpler solution would be to only drop packets larger then ethernet MTU (like most network drivers), but I guess that will break existing stuff already out there. The example below demonstrates what behaviour I'm referring to. Example: ---- 8< ------ # veth0 and veth1 is a veth pair and veth1 has ben moved to a separate network namespace. % # NS A % ip address add 1.2.3.4/24 dev veth0 % ip link set veth0 mtu 300 up % # NS B % ip address add 1.2.3.5/24 dev veth1 % ip link set veth1 mtu 1000 up % ping -c 1 -W 1 -s 400 1.2.3.4 PING 1.2.3.4 (1.2.3.4) 400(428) bytes of data. --- 1.2.3.4 ping statistics --- 1 packets transmitted, 0 received, 100% packet loss, time 0ms ---- 8< ------ While it works fine in most setup:s with hw-interfaces: ---- 8< ------ # Host A eth2 and Host B eth0 is on the same network. # On HOST A % ip address add 1.2.3.4/24 dev eth2 % ip link set eth2 mtu 300 up % # HOST B % ip address add 1.2.3.5/24 dev eth0 % ip link set eth0 mtu 1000 up % ping -c 1 -W 1 -s 400 1.2.3.4 PING 1.2.3.4 (1.2.3.4) 400(428) bytes of data. 408 bytes from 1.2.3.4: icmp_seq=1 ttl=64 time=1.57 ms --- 1.2.3.4 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 1.573/1.573/1.573/0.000 ms ---- 8< ------> > It is therefore untenable to remove this restriction univerally like > this.With the explaination above, do you still think it's untenable ? /Fredrik -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.linuxfoundation.org/pipermail/bridge/attachments/20170510/345490ba/attachment.html>
Fredrik Markström
2017-May-10 00:42 UTC
[Bridge] [PATCH 0/2] net: Set maximum receive packet size on veth interfaces
On Tue, May 9, 2017 at 7:48 PM, David Miller <davem at davemloft.net> wrote:> > From: Fredrik Markstrom <fredrik.markstrom at gmail.com> > Date: Tue, 9 May 2017 14:44:36 +0200 > > > Currently veth drops all packets larger then the mtu set on the > > receiving end of the pair. This is inconsistent with most hardware > > ethernet drivers. > > False. > > In fact, many pieces of ethernet hardware a not physically capable of > sending even VLAN packets that are above the normal base ethernet MTU. >Maybe I was unclear, the veth implementation drops all packers larger then the configured MTU (on the receiving interface). Most ethernet drivers accepts packets up to the ethernet MTU no matter the configured MTU. As far as I can tell from the RFC:s that is ok. A simpler solution would be to only drop packets larger then ethernet MTU (like most network drivers), but I guess that will break existing stuff already out there. The example below demonstrates what behaviour I'm referring to. Example: ---- 8< ------ # veth0 and veth1 is a veth pair and veth1 has ben moved to a separate network namespace. % # NS A % ip address add 1.2.3.4/24 dev veth0 % ip link set veth0 mtu 300 up % # NS B % ip address add 1.2.3.5/24 dev veth1 % ip link set veth1 mtu 1000 up % ping -c 1 -W 1 -s 400 1.2.3.4 PING 1.2.3.4 (1.2.3.4) 400(428) bytes of data. --- 1.2.3.4 ping statistics --- 1 packets transmitted, 0 received, 100% packet loss, time 0ms ---- 8< ------ While it works fine in most setup:s with hw-interfaces: ---- 8< ------ # Host A eth2 and Host B eth0 is on the same network. # On HOST A % ip address add 1.2.3.4/24 dev eth2 % ip link set eth2 mtu 300 up % # HOST B % ip address add 1.2.3.5/24 dev eth0 % ip link set eth0 mtu 1000 up % ping -c 1 -W 1 -s 400 1.2.3.4 PING 1.2.3.4 (1.2.3.4) 400(428) bytes of data. 408 bytes from 1.2.3.4: icmp_seq=1 ttl=64 time=1.57 ms --- 1.2.3.4 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 1.573/1.573/1.573/0.000 ms ---- 8< ------> > It is therefore untenable to remove this restriction univerally like > this.With the explaination above, do you still think it's untenable ? /Fredrik