search for: moongen

Displaying 12 results from an estimated 12 matches for "moongen".

Did you mean: monges
2017 Sep 27
2
[PATCH net-next RFC 5/5] vhost_net: basic tx virtqueue batched processing
...e batching on top. For zerocopy case and for simplicity, batched > > > processing were simply disabled by only fetching and processing one > > > descriptor at a time, this could be optimized in the future. > > > > > > XDP_DROP (without touching skb) on tun (with Moongen in guest) with > > > zercopy disabled: > > > > > > Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz: > > > Before: 3.20Mpps > > > After: 3.90Mpps (+22%) > > > > > > No differences were seen with zerocopy enabled. > > > > > &g...
2017 Sep 27
2
[PATCH net-next RFC 5/5] vhost_net: basic tx virtqueue batched processing
...e batching on top. For zerocopy case and for simplicity, batched > > > processing were simply disabled by only fetching and processing one > > > descriptor at a time, this could be optimized in the future. > > > > > > XDP_DROP (without touching skb) on tun (with Moongen in guest) with > > > zercopy disabled: > > > > > > Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz: > > > Before: 3.20Mpps > > > After: 3.90Mpps (+22%) > > > > > > No differences were seen with zerocopy enabled. > > > > > &g...
2017 Sep 27
2
[PATCH net-next RFC 0/5] batched tx processing in vhost_net
...sider the layout is re-designed completely, I don't see an easy method to reuse current 1.0 API for 1.1. > see comments on individual patches. I'm sure it can be addressed though. > >> Test shows about ~22% improvement in tx pss. > Is this with or without tx napi in guest? MoonGen is used in guest for better numbers. Thanks > >> Please review. >> >> Jason Wang (5): >> vhost: split out ring head fetching logic >> vhost: introduce helper to prefetch desc index >> vhost: introduce vhost_add_used_idx() >> vhost_net: rena...
2017 Sep 27
2
[PATCH net-next RFC 0/5] batched tx processing in vhost_net
...sider the layout is re-designed completely, I don't see an easy method to reuse current 1.0 API for 1.1. > see comments on individual patches. I'm sure it can be addressed though. > >> Test shows about ~22% improvement in tx pss. > Is this with or without tx napi in guest? MoonGen is used in guest for better numbers. Thanks > >> Please review. >> >> Jason Wang (5): >> vhost: split out ring head fetching logic >> vhost: introduce helper to prefetch desc index >> vhost: introduce vhost_add_used_idx() >> vhost_net: rena...
2017 Sep 26
2
[PATCH net-next RFC 5/5] vhost_net: basic tx virtqueue batched processing
...is also a requirement for doing > more batching on top. For zerocopy case and for simplicity, batched > processing were simply disabled by only fetching and processing one > descriptor at a time, this could be optimized in the future. > > XDP_DROP (without touching skb) on tun (with Moongen in guest) with > zercopy disabled: > > Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz: > Before: 3.20Mpps > After: 3.90Mpps (+22%) > > No differences were seen with zerocopy enabled. > > Signed-off-by: Jason Wang <jasowang at redhat.com> So where is the speedup comin...
2017 Sep 26
2
[PATCH net-next RFC 5/5] vhost_net: basic tx virtqueue batched processing
...is also a requirement for doing > more batching on top. For zerocopy case and for simplicity, batched > processing were simply disabled by only fetching and processing one > descriptor at a time, this could be optimized in the future. > > XDP_DROP (without touching skb) on tun (with Moongen in guest) with > zercopy disabled: > > Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz: > Before: 3.20Mpps > After: 3.90Mpps (+22%) > > No differences were seen with zerocopy enabled. > > Signed-off-by: Jason Wang <jasowang at redhat.com> So where is the speedup comin...
2017 Sep 27
0
[PATCH net-next RFC 5/5] vhost_net: basic tx virtqueue batched processing
...nt for doing >> more batching on top. For zerocopy case and for simplicity, batched >> processing were simply disabled by only fetching and processing one >> descriptor at a time, this could be optimized in the future. >> >> XDP_DROP (without touching skb) on tun (with Moongen in guest) with >> zercopy disabled: >> >> Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz: >> Before: 3.20Mpps >> After: 3.90Mpps (+22%) >> >> No differences were seen with zerocopy enabled. >> >> Signed-off-by: Jason Wang <jasowang at redhat.com&g...
2018 Jan 09
2
[PATCH net-next] vhost_net: batch used ring update in rx
...ion for both batch dequeuing and dpdk batching, VHOST_RX_BATCH was used as the maximum number of descriptors that could be batched. Test were done between two machines with 2.40GHz Intel(R) Xeon(R) CPU E5-2630 connected back to back through ixgbe. Traffic were generated on one remote ixgbe through MoonGen and measure the RX pps through testpmd in guest when do xdp_redirect_map from local ixgbe to tap. RX pps were increased from 3.05 Mpps to 4.00 Mpps (about 31% improvement). One possible concern for this is the implications for TCP (especially latency sensitive workload). Result[1] does not show ob...
2017 Sep 27
0
[PATCH net-next RFC 0/5] batched tx processing in vhost_net
...get buffers then you use them. It is not tied to actual separate used ring. > > see comments on individual patches. I'm sure it can be addressed though. > > > > > Test shows about ~22% improvement in tx pss. > > Is this with or without tx napi in guest? > > MoonGen is used in guest for better numbers. > > Thanks Not sure I understand. Did you set napi_tx to true or false? > > > > > Please review. > > > > > > Jason Wang (5): > > > vhost: split out ring head fetching logic > > > vhost: introduc...
2017 Sep 22
0
[PATCH net-next RFC 5/5] vhost_net: basic tx virtqueue batched processing
...d updating used ring. It is also a requirement for doing more batching on top. For zerocopy case and for simplicity, batched processing were simply disabled by only fetching and processing one descriptor at a time, this could be optimized in the future. XDP_DROP (without touching skb) on tun (with Moongen in guest) with zercopy disabled: Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz: Before: 3.20Mpps After: 3.90Mpps (+22%) No differences were seen with zerocopy enabled. Signed-off-by: Jason Wang <jasowang at redhat.com> --- drivers/vhost/net.c | 215 ++++++++++++++++++++++++++++--------------...
2017 Sep 22
17
[PATCH net-next RFC 0/5] batched tx processing in vhost_net
Hi: This series tries to implement basic tx batched processing. This is done by prefetching descriptor indices and update used ring in a batch. This intends to speed up used ring updating and improve the cache utilization. Test shows about ~22% improvement in tx pss. Please review. Jason Wang (5): vhost: split out ring head fetching logic vhost: introduce helper to prefetch desc index
2017 Sep 22
17
[PATCH net-next RFC 0/5] batched tx processing in vhost_net
Hi: This series tries to implement basic tx batched processing. This is done by prefetching descriptor indices and update used ring in a batch. This intends to speed up used ring updating and improve the cache utilization. Test shows about ~22% improvement in tx pss. Please review. Jason Wang (5): vhost: split out ring head fetching logic vhost: introduce helper to prefetch desc index