Displaying 20 results from an estimated 300 matches similar to: "[Bridge] NIC unicast macs table manipulation by bridge"
2007 Apr 18
2
[Bridge] tg3 bridge problems
Hello,
I've got a very strange problem. Lately I've been setting up my linux
servers for network (layer2) redundancy with a bridge interface containing
two ethernet interfaces connecting to two switches. So far I didn't have
any problems with it, but now a very strange thing happens with a new
server I'm installing. The server is an ibm x346 having two onboard
BCM5721 cards, the
2015 Oct 28
2
[PATCH v2 1/3] virtio_net: Stop doing DMA from the stack
From: Andy Lutomirski <luto at amacapital.net>
Once virtio starts using the DMA API, we won't be able to safely DMA
from the stack. virtio-net does a couple of config DMA requests
from small stack buffers -- switch to using dynamically-allocated
memory.
This should have no effect on any performance-critical code paths.
Cc: netdev at vger.kernel.org
Cc: "Michael S. Tsirkin"
2015 Oct 28
2
[PATCH v2 1/3] virtio_net: Stop doing DMA from the stack
From: Andy Lutomirski <luto at amacapital.net>
Once virtio starts using the DMA API, we won't be able to safely DMA
from the stack. virtio-net does a couple of config DMA requests
from small stack buffers -- switch to using dynamically-allocated
memory.
This should have no effect on any performance-critical code paths.
Cc: netdev at vger.kernel.org
Cc: "Michael S. Tsirkin"
2009 Aug 05
2
bridge vs macvlan performance (was: some veth related issues)
Ben Greear wrote:
> Well, it seems we could and should fix veth to work, but it will have
> to do equivalent work of copying an skb most likely, so either way
> you'll probably get a big performance hit.
Using the same pktgen script (i.e with clone=0) I see that a
veth-->bridge-->veth configuration gives about 400K PPS forwarding
performance where
2009 Aug 05
2
bridge vs macvlan performance (was: some veth related issues)
Ben Greear wrote:
> Well, it seems we could and should fix veth to work, but it will have
> to do equivalent work of copying an skb most likely, so either way
> you'll probably get a big performance hit.
Using the same pktgen script (i.e with clone=0) I see that a
veth-->bridge-->veth configuration gives about 400K PPS forwarding
performance where
2007 Apr 18
4
[Bridge] Is bridgeing possible with kernel 2.6.10
Hi
I didn't get bridgeing to work with Linux kernel 2.6.10 (coldfire 5485)
yet. When i try to configure bridgeing
with "brctl", everything seems to be ok, but i if i try to ping through
my bridge, i see ARP-Request arriving on the other side, but the
ARP-Responses do not go back through the bridge.
Now i could read in
2018 Apr 19
1
[PATCH v2 net 1/3] virtio_net: split out ctrl buffer
When sending control commands, virtio net sets up several buffers for
DMA. The buffers are all part of the net device which means it's
actually allocated by kvmalloc so it's in theory (on extreme memory
pressure) possible to get a vmalloc'ed buffer which on some platforms
means we can't DMA there.
Fix up by moving the DMA buffers into a separate structure.
Reported-by: Mikulas
2018 Apr 19
3
[PATCH net] virtio_net: split out ctrl buffer
When sending control commands, virtio net sets up several buffers for
DMA. The buffers are all part of the net device which means it's
actually allocated by kvmalloc so in theory (on extreme memory pressure)
it's possible to get a vmalloc'ed buffer which on some platforms means
we can't DMA there.
Fix up by moving the DMA buffers out into a separate structure.
Reported-by:
2008 Jul 12
4
[PATCH] tun: Fix/rewrite packet filtering logic
Please see the following thread to get some context on this
http://marc.info/?l=linux-netdev&m=121564433018903&w=2
Basically the issue is that current multi-cast filtering stuff in
the TUN/TAP driver is seriously broken.
Original patch went in without proper review and ACK. It was broken and
confusing to start with and subsequent patches broke it completely.
To give you an idea of
2008 Jul 12
4
[PATCH] tun: Fix/rewrite packet filtering logic
Please see the following thread to get some context on this
http://marc.info/?l=linux-netdev&m=121564433018903&w=2
Basically the issue is that current multi-cast filtering stuff in
the TUN/TAP driver is seriously broken.
Original patch went in without proper review and ACK. It was broken and
confusing to start with and subsequent patches broke it completely.
To give you an idea of
2009 Nov 02
0
[PATCHv4 3/6] qemu/net: add raw backend
Add raw network backend option which uses a packet socket to provide
raw networking access. Once the socket is opened it's bound to a
provided host interface, such that packets received on the interface
are delivered to the VM and packets sent by the VM are sent to the
interface.
This is functionally similar to the existing pcap network
backend, with the same advantages and problems.
2009 Nov 02
0
[PATCHv4 3/6] qemu/net: add raw backend
Add raw network backend option which uses a packet socket to provide
raw networking access. Once the socket is opened it's bound to a
provided host interface, such that packets received on the interface
are delivered to the VM and packets sent by the VM are sent to the
interface.
This is functionally similar to the existing pcap network
backend, with the same advantages and problems.
2015 Oct 28
10
[PATCH v3 0/3] virtio DMA API core stuff
This switches virtio to use the DMA API unconditionally. I'm sure
it breaks things, but it seems to work on x86 using virtio-pci, with
and without Xen, and using both the modern 1.0 variant and the
legacy variant.
Changes from v2:
- Fix really embarrassing bug. This version actually works.
Changes from v1:
- Fix an endian conversion error causing a BUG to hit.
- Fix a DMA ordering issue
2015 Oct 28
10
[PATCH v3 0/3] virtio DMA API core stuff
This switches virtio to use the DMA API unconditionally. I'm sure
it breaks things, but it seems to work on x86 using virtio-pci, with
and without Xen, and using both the modern 1.0 variant and the
legacy variant.
Changes from v2:
- Fix really embarrassing bug. This version actually works.
Changes from v1:
- Fix an endian conversion error causing a BUG to hit.
- Fix a DMA ordering issue
2009 Sep 16
4
vhost-net todo list
Some people asked about getting involved with vhost.
Here's a short list of projects.
vhost-net driver projects
- profiling would be very helpful, I have not done any yet
- tap support - working on it now
- merged buffers - working on it now
- scalability/fairness for large # of guests - working on it now
- logging support with dirty page tracking in kernel - working on it now
- indirect
2009 Sep 16
4
vhost-net todo list
Some people asked about getting involved with vhost.
Here's a short list of projects.
vhost-net driver projects
- profiling would be very helpful, I have not done any yet
- tap support - working on it now
- merged buffers - working on it now
- scalability/fairness for large # of guests - working on it now
- logging support with dirty page tracking in kernel - working on it now
- indirect
2023 May 24
1
[PATCH V3 net-next 1/2] virtio-net: convert rx mode setting to use workqueue
This patch convert rx mode setting to be done in a workqueue, this is
a must for allow to sleep when waiting for the cvq command to
response since current code is executed under addr spin lock.
Signed-off-by: Jason Wang <jasowang at redhat.com>
---
Changes since V1:
- use RTNL to synchronize rx mode worker
---
drivers/net/virtio_net.c | 55 +++++++++++++++++++++++++++++++++++++---
1 file
2009 Mar 16
0
[Bridge] STP FD when STP is disabled
It appears that interfaces participate in STP's forwarding delay even when STP is disabled for a given bridge. I can remedy by setting the FD to 0 for each new bridge, but I'm wondering if this is by design or oversight.
[root@ ~]# brctl show
bridge name bridge id STP enabled interfaces
br0.690 8000.00065b8bce3e no eth1.690
2023 May 24
2
[PATCH V3 net-next 0/2] virtio-net: don't busy poll for cvq command
Hi all:
The code used to busy poll for cvq command which turns out to have
several side effects:
1) infinite poll for buggy devices
2) bad interaction with scheduler
So this series tries to use cond_resched() in the waiting loop. Before
doing this we need first make sure the cvq command is not executed in
atomic environment, so we need first convert rx mode handling to a
workqueue.
Please
2023 Jul 20
2
[PATCH net-next v4 0/2] virtio-net: don't busy poll for cvq command
Hi all:
The code used to busy poll for cvq command which turns out to have
several side effects:
1) infinite poll for buggy devices
2) bad interaction with scheduler
So this series tries to use cond_resched() in the waiting loop. Before
doing this we need first make sure the cvq command is not executed in
atomic environment, so we need first convert rx mode handling to a
workqueue.
Note that,