similar to: [net-next, v6, 6/7] net-sysfs: Add interface for Rx queue(s) map per Tx queue

Displaying 20 results from an estimated 400 matches similar to: "[net-next, v6, 6/7] net-sysfs: Add interface for Rx queue(s) map per Tx queue"

2018 Aug 03
0
[net-next, v6, 6/7] net-sysfs: Add interface for Rx queue(s) map per Tx queue
On Fri, Aug 03, 2018 at 12:06:51PM -0700, Andrei Vagin wrote: > On Fri, Aug 03, 2018 at 12:08:05AM +0300, Michael S. Tsirkin wrote: > > On Thu, Aug 02, 2018 at 02:04:12PM -0700, Nambiar, Amritha wrote: > > > On 8/1/2018 5:11 PM, Andrei Vagin wrote: > > > > On Tue, Jul 10, 2018 at 07:28:49PM -0700, Nambiar, Amritha wrote: > > > >> With this patch series,
2018 Aug 07
1
[PATCH net-next] net: allow to call netif_reset_xps_queues() under cpu_read_lock
From: Andrei Vagin <avagin at gmail.com> The definition of static_key_slow_inc() has cpus_read_lock in place. In the virtio_net driver, XPS queues are initialized after setting the queue:cpu affinity in virtnet_set_affinity() which is already protected within cpus_read_lock. Lockdep prints a warning when we are trying to acquire cpus_read_lock when it is already held. This patch adds an
2018 Jul 26
2
net-next boot error
On Thu, Jul 26, 2018 at 11:29 AM, syzbot <syzbot+604f8271211546f5b3c7 at syzkaller.appspotmail.com> wrote: > Hello, > > syzbot found the following crash on: > > HEAD commit: dc66fe43b7eb rds: send: Fix dead code in rds_sendmsg > git tree: net-next > console output: https://syzkaller.appspot.com/x/log.txt?x=127874c8400000 > kernel config:
2018 Jul 26
2
net-next boot error
On Thu, Jul 26, 2018 at 11:29 AM, syzbot <syzbot+604f8271211546f5b3c7 at syzkaller.appspotmail.com> wrote: > Hello, > > syzbot found the following crash on: > > HEAD commit: dc66fe43b7eb rds: send: Fix dead code in rds_sendmsg > git tree: net-next > console output: https://syzkaller.appspot.com/x/log.txt?x=127874c8400000 > kernel config:
2018 Jul 26
0
net-next boot error
[ Added Thomas Gleixner ] On Thu, 26 Jul 2018 11:34:39 +0200 Dmitry Vyukov <dvyukov at google.com> wrote: > On Thu, Jul 26, 2018 at 11:29 AM, syzbot > <syzbot+604f8271211546f5b3c7 at syzkaller.appspotmail.com> wrote: > > Hello, > > > > syzbot found the following crash on: > > > > HEAD commit: dc66fe43b7eb rds: send: Fix dead code in rds_sendmsg
2013 Nov 05
1
[RESEND PATCH net-next] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing user to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support,
2013 Nov 05
1
[RESEND PATCH net-next] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing user to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support,
2016 Aug 12
0
[PATCH 6/6] net: virtio-net: Convert to hotplug state machine
Install the callbacks via the state machine. The driver supports multiple instances and therefore the new cpuhp_state_add_instance_nocalls() infrastrucure is used. The driver currently uses get_online_cpus() to avoid missing a CPU hotplug event while invoking virtnet_set_affinity(). This could be avoided by using cpuhp_state_add_instance() variant which holds the hotplug lock and invokes callback
2016 Aug 12
0
[PATCH 6/6] net: virtio-net: Convert to hotplug state machine
Install the callbacks via the state machine. The driver supports multiple instances and therefore the new cpuhp_state_add_instance_nocalls() infrastrucure is used. The driver currently uses get_online_cpus() to avoid missing a CPU hotplug event while invoking virtnet_set_affinity(). This could be avoided by using cpuhp_state_add_instance() variant which holds the hotplug lock and invokes callback
2013 Sep 29
0
[PATCH net-next] virtio-net: switch to use XPS to choose txq
Jason Wang <jasowang at redhat.com> writes: > We used to use a percpu structure vq_index to record the cpu to queue > mapping, this is suboptimal since it duplicates the work of XPS and > loses all other XPS functionality such as allowing use to configure > their own transmission steering strategy. > > So this patch switches to use XPS and suggest a default mapping when
2013 Sep 30
0
[net-next PATCH V2] virtio-net: switch to use XPS to choose txq
On Mon, Sep 30, 2013 at 03:37:17PM +0800, Jason Wang wrote: > We used to use a percpu structure vq_index to record the cpu to queue > mapping, this is suboptimal since it duplicates the work of XPS and > loses all other XPS functionality such as allowing use to configure > their own transmission steering strategy. > > So this patch switches to use XPS and suggest a default
2013 Sep 27
0
[PATCH net-next] virtio-net: switch to use XPS to choose txq
On Fri, Sep 27, 2013 at 01:57:24PM +0800, Jason Wang wrote: > We used to use a percpu structure vq_index to record the cpu to queue > mapping, this is suboptimal since it duplicates the work of XPS and > loses all other XPS functionality such as allowing use to configure > their own transmission steering strategy. > > So this patch switches to use XPS and suggest a default
2013 Sep 30
4
[net-next PATCH V2] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing use to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support, there's
2013 Sep 30
4
[net-next PATCH V2] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing use to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support, there's
2019 Dec 23
0
[RHEL7]virtio-net: switch to use XPS to choose txq
On 2019/12/20 ??7:48, Cindylu wrote: > Bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=1769479 > Upstream Status: 9bb8ca86075f37d3c169b9c46f8e7c6d3165e18f > Brew: ?https://brewweb.devel.redhat.com/taskinfo?taskID=25482050 > Tested: verified by customer > > commit 9bb8ca86075f37d3c169b9c46f8e7c6d3165e18f > Author: Jason Wang <jasowang at redhat.com> > Date: Tue
2013 Oct 01
0
[net-next PATCH V2] virtio-net: switch to use XPS to choose txq
From: Jason Wang <jasowang at redhat.com> Date: Mon, 30 Sep 2013 15:37:17 +0800 > We used to use a percpu structure vq_index to record the cpu to queue > mapping, this is suboptimal since it duplicates the work of XPS and > loses all other XPS functionality such as allowing use to configure > their own transmission steering strategy. > > So this patch switches to use XPS
2013 Sep 27
4
[PATCH net-next] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing use to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support, there's
2013 Sep 27
4
[PATCH net-next] virtio-net: switch to use XPS to choose txq
We used to use a percpu structure vq_index to record the cpu to queue mapping, this is suboptimal since it duplicates the work of XPS and loses all other XPS functionality such as allowing use to configure their own transmission steering strategy. So this patch switches to use XPS and suggest a default mapping when the number of cpus is equal to the number of queues. With XPS support, there's
2014 Jan 16
2
[PATCH net-next v3 4/5] net-sysfs: add support for device-specific rx queue sysfs attributes
On Thu, 2014-01-16 at 01:38 -0800, Michael Dalton wrote: [...] > --- a/include/linux/netdevice.h > +++ b/include/linux/netdevice.h [...] > @@ -2401,6 +2416,23 @@ static inline int netif_copy_real_num_queues(struct net_device *to_dev, > #endif > } > > +#ifdef CONFIG_SYSFS > +static inline unsigned int get_netdev_rx_queue_index( > + struct netdev_rx_queue *queue) >
2014 Jan 16
0
[PATCH net-next v3 4/5] net-sysfs: add support for device-specific rx queue sysfs attributes
On Jan 16, 2014 at 10:57 AM, Ben Hutchings <bhutchings at solarflare.com> wrote: > Why write a loop when you can do: > i = queue - dev->_rx; Good point, the loop approach was done in get_netdev_queue_index -- I agree your fix is faster and simpler. I'll fix in next patchset. Thanks! Best, Mike