Displaying 6 results from an estimated 6 matches for "process_buff".
Did you mean:
process_buffer
2013 Jun 08
7
[virtio-spec PATCH 0/5] Receiving Used Buffers example code: cleanups and an extra mb()
Patches before the last are small cleanups.
In the last patch I'm trying to extract / generalize an idea from Stefan
Hajnoczi's review of my virtio-net driver for OVMF.
Unfortunately I can't find Stefan's email on any mailing list archive
(sourceforge, gmane, mail-archive etc. all have only my response), so
I'll quote it here.
The patch Stefan was reviewing is
2013 Jun 08
7
[virtio-spec PATCH 0/5] Receiving Used Buffers example code: cleanups and an extra mb()
Patches before the last are small cleanups.
In the last patch I'm trying to extract / generalize an idea from Stefan
Hajnoczi's review of my virtio-net driver for OVMF.
Unfortunately I can't find Stefan's email on any mailing list archive
(sourceforge, gmane, mail-archive etc. all have only my response), so
I'll quote it here.
The patch Stefan was reviewing is
2018 Dec 17
1
kernel vhost demands an interrupt from guest when the ring is full in order to enable guest to submit new packets to the queue
...) {
virtq_enable_used_buffer_notifications(vq);
mb();
if (vq->last_seen_used != le16_to_cpu(virtq->used.idx))
break;
virtq_disable_used_buffer_notifications(vq);
}
struct virtq_used_elem *e = virtq.used->ring[vq->last_seen_used%vsz];
process_buffer(e);
vq->last_seen_used++;
>
>
> Steven
>
> _______________________________________________
> Virtualization mailing list
> Virtualization at lists.linux-foundation.org
> https://lists.linuxfoundation.org/mail...
2018 Dec 13
2
kernel vhost demands an interrupt from guest when the ring is full in order to enable guest to submit new packets to the queue
Folks,
We came across a memory race condition between VPP vhost driver and the kernel vhost. VPP is running a tap interface over vhost backend. In this case, VPP is acting as the vhost driver mode and the kernel vhost is acting as the vhost device mode.
In the kernel vhost?s TX traffic direction which is VPP?s RX traffic direction, kernel vhost is the producer and VPP is the consumer. Kernel
2018 Dec 13
2
kernel vhost demands an interrupt from guest when the ring is full in order to enable guest to submit new packets to the queue
Folks,
We came across a memory race condition between VPP vhost driver and the kernel vhost. VPP is running a tap interface over vhost backend. In this case, VPP is acting as the vhost driver mode and the kernel vhost is acting as the vhost device mode.
In the kernel vhost?s TX traffic direction which is VPP?s RX traffic direction, kernel vhost is the producer and VPP is the consumer. Kernel
2018 Dec 17
0
kernel vhost demands an interrupt from guest when the ring is full in order to enable guest to submit new packets to the queue
...le16_to_cpu(virtq->used.idx)) {
virtq_enable_used_buffer_notifications(vq);
mb();
if (vq->last_seen_used != le16_to_cpu(virtq->used.idx))
break;
virtq_disable_used_buffer_notifications(vq);
}
struct virtq_used_elem *e = virtq.used->ring[vq->last_seen_used%vsz];
process_buffer(e);
vq->last_seen_used++;
>
>
> Steven
>
> _______________________________________________
> Virtualization mailing list
> Virtualization at lists.linux-foundation.org
> https://lists.linuxfoundation.org/mailman/listinfo/virtualization