similar to: No subject

Displaying 20 results from an estimated 700 matches similar to: "No subject"

2017 Dec 07
2
[PATCH net-next] virtio_net: Disable interrupts if napi_complete_done rescheduled napi
Since commit 39e6c8208d7b ("net: solve a NAPI race") napi has been able to be rescheduled within napi_complete_done() even in non-busypoll case, but virtnet_poll() always enabled interrupts before complete, and when napi was rescheduled within napi_complete_done() it did not disable interrupts. This caused more interrupts when event idx is disabled. According to commit cbdadbbf0c79
2017 Dec 07
2
[PATCH net-next] virtio_net: Disable interrupts if napi_complete_done rescheduled napi
Since commit 39e6c8208d7b ("net: solve a NAPI race") napi has been able to be rescheduled within napi_complete_done() even in non-busypoll case, but virtnet_poll() always enabled interrupts before complete, and when napi was rescheduled within napi_complete_done() it did not disable interrupts. This caused more interrupts when event idx is disabled. According to commit cbdadbbf0c79
2017 Dec 07
0
[PATCH net-next] virtio_net: Disable interrupts if napi_complete_done rescheduled napi
On Thu, Dec 07, 2017 at 01:15:15PM +0900, Toshiaki Makita wrote: > Since commit 39e6c8208d7b ("net: solve a NAPI race") napi has been able > to be rescheduled within napi_complete_done() even in non-busypoll case, > but virtnet_poll() always enabled interrupts before complete, and when > napi was rescheduled within napi_complete_done() it did not disable > interrupts. >
2020 Jun 11
0
[PATCH RFC v7 03/14] vhost: use batched get_vq_desc version
On Wed, Jun 10, 2020 at 06:18:32PM +0200, Eugenio Perez Martin wrote: > On Wed, Jun 10, 2020 at 5:13 PM Michael S. Tsirkin <mst at redhat.com> wrote: > > > > On Wed, Jun 10, 2020 at 02:37:50PM +0200, Eugenio Perez Martin wrote: > > > > +/* This function returns a value > 0 if a descriptor was found, or 0 if none were found. > > > > + * A negative
2017 Dec 07
1
[PATCH net-next] virtio_net: Disable interrupts if napi_complete_done rescheduled napi
On 2017?12?07? 13:09, Michael S. Tsirkin wrote: > On Thu, Dec 07, 2017 at 01:15:15PM +0900, Toshiaki Makita wrote: >> Since commit 39e6c8208d7b ("net: solve a NAPI race") napi has been able >> to be rescheduled within napi_complete_done() even in non-busypoll case, >> but virtnet_poll() always enabled interrupts before complete, and when >> napi was rescheduled
2017 Dec 07
1
[PATCH net-next] virtio_net: Disable interrupts if napi_complete_done rescheduled napi
On 2017?12?07? 13:09, Michael S. Tsirkin wrote: > On Thu, Dec 07, 2017 at 01:15:15PM +0900, Toshiaki Makita wrote: >> Since commit 39e6c8208d7b ("net: solve a NAPI race") napi has been able >> to be rescheduled within napi_complete_done() even in non-busypoll case, >> but virtnet_poll() always enabled interrupts before complete, and when >> napi was rescheduled
2009 Sep 04
2
Xen & netperf
First, I apologize if this message has been received multiple times. I''m having problems subscribing to this mailing list: Hi xen-users, I am trying to decide whether I should run a game server inside a Xen domain. My primary reason for wanting to virtualize is because I want to isolate this environment from the rest of my server. I really like the idea of isolating the game server
2020 Jun 16
0
[PATCH RFC v7 03/14] vhost: use batched get_vq_desc version
On Tue, Jun 16, 2020 at 05:23:43PM +0200, Eugenio Perez Martin wrote: > On Mon, Jun 15, 2020 at 6:05 PM Eugenio P??rez <eperezma at redhat.com> wrote: > > > > On Thu, 2020-06-11 at 07:30 -0400, Michael S. Tsirkin wrote: > > > On Wed, Jun 10, 2020 at 06:18:32PM +0200, Eugenio Perez Martin wrote: > > > > On Wed, Jun 10, 2020 at 5:13 PM Michael S. Tsirkin
2009 Oct 26
2
performance regression in virtio-net in 2.6.32-rc4
Hi! I noticed a performance regression in virtio net: going from 2.6.31 to 2.6.32-rc4 I see this, for guest to host communication: [mst at tuck ~]$ ssh robin sh streamtest1 TCP STREAM TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 11.0.0.3 (11.0.0.3) port 0 AF_INET : demo Recv Send Send Socket Socket Message Elapsed Size Size Size Time Throughput bytes bytes bytes secs.
2009 Oct 26
2
performance regression in virtio-net in 2.6.32-rc4
Hi! I noticed a performance regression in virtio net: going from 2.6.31 to 2.6.32-rc4 I see this, for guest to host communication: [mst at tuck ~]$ ssh robin sh streamtest1 TCP STREAM TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 11.0.0.3 (11.0.0.3) port 0 AF_INET : demo Recv Send Send Socket Socket Message Elapsed Size Size Size Time Throughput bytes bytes bytes secs.
2013 Oct 31
0
[PATCH net-next 2/2] virtio-net: coalesce rx frags when possible during rx
Commit 2613af0ed18a11d5c566a81f9a6510b73180660a (virtio_net: migrate mergeable rx buffers to page frag allocators) try to increase the payload/truesize for MTU-sized traffic. But this will introduce the extra overhead for GSO packets received because of the frag list. This commit tries to reduce this issue by coalesce the possible rx frags when possible during rx. Test result shows the about 15%
2013 Oct 31
0
[PATCH net-next V2 2/2] virtio-net: coalesce rx frags when possible during rx
Commit 2613af0ed18a11d5c566a81f9a6510b73180660a (virtio_net: migrate mergeable rx buffers to page frag allocators) try to increase the payload/truesize for MTU-sized traffic. But this will introduce the extra overhead for GSO packets received because of the frag list. This commit tries to reduce this issue by coalesce the possible rx frags when possible during rx. Test result shows the about 15%
2013 Nov 01
0
[PATCH net-next V3 2/2] virtio-net: coalesce rx frags when possible during rx
Commit 2613af0ed18a11d5c566a81f9a6510b73180660a (virtio_net: migrate mergeable rx buffers to page frag allocators) try to increase the payload/truesize for MTU-sized traffic. But this will introduce the extra overhead for GSO packets received because of the frag list. This commit tries to reduce this issue by coalesce the possible rx frags when possible during rx. Test result shows the about 15%
2011 Nov 29
1
[PATCH] virtio-ring: Use threshold for switching to indirect descriptors
Currently if VIRTIO_RING_F_INDIRECT_DESC is enabled we will use indirect descriptors even if we have plenty of space in the ring. This means that we take a performance hit at all times due to the overhead of creating indirect descriptors. With this patch, we will use indirect descriptors only if we have less than either 16, or 12% of the total amount of descriptors available. I did basic
2011 Nov 29
1
[PATCH] virtio-ring: Use threshold for switching to indirect descriptors
Currently if VIRTIO_RING_F_INDIRECT_DESC is enabled we will use indirect descriptors even if we have plenty of space in the ring. This means that we take a performance hit at all times due to the overhead of creating indirect descriptors. With this patch, we will use indirect descriptors only if we have less than either 16, or 12% of the total amount of descriptors available. I did basic
2020 Jul 01
0
[PATCH RFC v8 02/11] vhost: use batched get_vq_desc version
On 2020/7/1 ??9:04, Eugenio Perez Martin wrote: > On Wed, Jul 1, 2020 at 2:40 PM Jason Wang <jasowang at redhat.com> wrote: >> >> On 2020/7/1 ??6:43, Eugenio Perez Martin wrote: >>> On Tue, Jun 23, 2020 at 6:15 PM Eugenio Perez Martin >>> <eperezma at redhat.com> wrote: >>>> On Mon, Jun 22, 2020 at 6:29 PM Michael S. Tsirkin <mst at
2020 Jul 09
0
[PATCH RFC v8 02/11] vhost: use batched get_vq_desc version
On Thu, Jul 09, 2020 at 06:46:13PM +0200, Eugenio Perez Martin wrote: > On Wed, Jul 1, 2020 at 4:10 PM Jason Wang <jasowang at redhat.com> wrote: > > > > > > On 2020/7/1 ??9:04, Eugenio Perez Martin wrote: > > > On Wed, Jul 1, 2020 at 2:40 PM Jason Wang <jasowang at redhat.com> wrote: > > >> > > >> On 2020/7/1 ??6:43, Eugenio Perez
2020 Jul 01
0
[PATCH RFC v8 02/11] vhost: use batched get_vq_desc version
On 2020/7/1 ??6:43, Eugenio Perez Martin wrote: > On Tue, Jun 23, 2020 at 6:15 PM Eugenio Perez Martin > <eperezma at redhat.com> wrote: >> On Mon, Jun 22, 2020 at 6:29 PM Michael S. Tsirkin <mst at redhat.com> wrote: >>> On Mon, Jun 22, 2020 at 06:11:21PM +0200, Eugenio Perez Martin wrote: >>>> On Mon, Jun 22, 2020 at 5:55 PM Michael S. Tsirkin <mst
2017 Apr 24
8
[PATCH net-next v3 0/5] virtio-net tx napi
From: Willem de Bruijn <willemb at google.com> Add napi for virtio-net transmit completion processing. Changes: v2 -> v3: - convert __netif_tx_trylock to __netif_tx_lock on tx napi poll ensure that the handler always cleans, to avoid deadlock - unconditionally clean in start_xmit avoid adding an unnecessary "if (use_napi)" branch - remove
2017 Apr 24
8
[PATCH net-next v3 0/5] virtio-net tx napi
From: Willem de Bruijn <willemb at google.com> Add napi for virtio-net transmit completion processing. Changes: v2 -> v3: - convert __netif_tx_trylock to __netif_tx_lock on tx napi poll ensure that the handler always cleans, to avoid deadlock - unconditionally clean in start_xmit avoid adding an unnecessary "if (use_napi)" branch - remove