similar to: [vhost] sending data to guest, async

Displaying 20 results from an estimated 20000 matches similar to: "[vhost] sending data to guest, async"

2007 Dec 14
0
async-observer rails plugin
I''m pleased to announce the existence of async-observer. This is the very first public release. In the future I''ll confine announcements to the beanstalk-talk mailing list. WHAT IS ASYNC OBSERVER? ----------------------- Async Observer is a Rails plugin that provides deep integration with Beanstalk. beanstalkd is a fast, distributed, in-memory work-queue service. Its
2015 May 19
1
Trying to make VHOST-BLK to work
Hi Kernel KVM folks and Michael, I am trying to make VHOST-BLK to work by myself since I did not hear from Asias on my personal mail to him. I will be grateful to you for your time and response. The problem I am seeing is there is no response seen by the VM through the HOST has executed the IO successfully and issued vhost_add_used_and_signal. Here is my question. Which code in QEMU issues
2015 May 19
1
Trying to make VHOST-BLK to work
Hi Kernel KVM folks and Michael, I am trying to make VHOST-BLK to work by myself since I did not hear from Asias on my personal mail to him. I will be grateful to you for your time and response. The problem I am seeing is there is no response seen by the VM through the HOST has executed the IO successfully and issued vhost_add_used_and_signal. Here is my question. Which code in QEMU issues
2024 May 28
1
Async Query Directory (for VFS)
Hi, is there a branch, patch or other attempts I can look at for async processing of SMB2 QUERY_DIRECTORY? Looking online I see proposals from years ago for making everything async but I don't know what happened to it? In my use case, I don't care as much about impersonation and security as it's being used for client only loopback on Mac, process running as the logged in user. I have
2024 May 28
1
Async Query Directory (for VFS)
On 5/28/24 10:06 AM, Dan Greenfield via samba wrote: > Hi, is there a branch, patch or other attempts I can look at for > async processing of SMB2 QUERY_DIRECTORY? Looking online I see > proposals from years ago for making everything async but I don't know > what happened to it? probably the same as to many other WIP branches: sitting there getting dusty. :) > In my use case,
2024 May 28
1
Async Query Directory (for VFS)
Thanks Ralph. Unfortunately that option doesn?t quite solve it. However even if it did, it?d still be good to add Async support for such queries. If there is WIP then I can look into finishing off the code myself. Can you point me to any branch that has this WIP? Looking through Samba source3 internals, it looks somewhat straightforward to add primitive support along the lines of Async pread,
2019 Apr 04
1
[PATCH RFC 2/4] vhost/vsock: split packets to send using multiple buffers
If the packets to sent to the guest are bigger than the buffer available, we can split them, using multiple buffers and fixing the length in the packet header. This is safe since virtio-vsock supports only stream sockets. Signed-off-by: Stefano Garzarella <sgarzare at redhat.com> --- drivers/vhost/vsock.c | 35 +++++++++++++++++++++++++++++------ 1 file changed, 29 insertions(+), 6
2019 Apr 05
0
[PATCH RFC 2/4] vhost/vsock: split packets to send using multiple buffers
On Thu, Apr 04, 2019 at 12:58:36PM +0200, Stefano Garzarella wrote: > @@ -139,8 +139,18 @@ vhost_transport_do_send_pkt(struct vhost_vsock *vsock, > break; > } > > - len = iov_length(&vq->iov[out], in); > - iov_iter_init(&iov_iter, READ, &vq->iov[out], in, len); > + payload_len = pkt->len - pkt->off; > + iov_len =
2008 Sep 08
2
Problems with async worker request
Sorry if this comes through twice. I already sent this once, before joining the mailing list. I''m attempting to use Backgroundrb to handle asynchronous pdf creation, but in doing so, I''ve run into a very strange problem. Below is a method that''s called from the controller which creates a new worker, then grabs the worker and calls the ''build_pdf''
2019 Apr 05
1
[PATCH RFC 2/4] vhost/vsock: split packets to send using multiple buffers
On Fri, Apr 05, 2019 at 09:13:56AM +0100, Stefan Hajnoczi wrote: > On Thu, Apr 04, 2019 at 12:58:36PM +0200, Stefano Garzarella wrote: > > @@ -139,8 +139,18 @@ vhost_transport_do_send_pkt(struct vhost_vsock *vsock, > > break; > > } > > > > - len = iov_length(&vq->iov[out], in); > > - iov_iter_init(&iov_iter, READ, &vq->iov[out],
2019 Apr 05
1
[PATCH RFC 2/4] vhost/vsock: split packets to send using multiple buffers
On Fri, Apr 05, 2019 at 09:13:56AM +0100, Stefan Hajnoczi wrote: > On Thu, Apr 04, 2019 at 12:58:36PM +0200, Stefano Garzarella wrote: > > @@ -139,8 +139,18 @@ vhost_transport_do_send_pkt(struct vhost_vsock *vsock, > > break; > > } > > > > - len = iov_length(&vq->iov[out], in); > > - iov_iter_init(&iov_iter, READ, &vq->iov[out],
2019 Jul 17
0
[PATCH v4 4/5] vhost/vsock: split packets to send using multiple buffers
If the packets to sent to the guest are bigger than the buffer available, we can split them, using multiple buffers and fixing the length in the packet header. This is safe since virtio-vsock supports only stream sockets. Signed-off-by: Stefano Garzarella <sgarzare at redhat.com> --- drivers/vhost/vsock.c | 66 ++++++++++++++++++-------
2012 Sep 05
6
Async AGI
Hi, Is there a way to execute next priority in the dialplan if you have called agi:async? I want to play warning message if adhearsion is down. Currently I wasn't able to make it work. The dialplan execution ends after the first priority. [incomming] exten => _X.,1,AGI(agi:async) exten => _X.,2,Answer exten => _X.,3,Playback(some-message) exten => _X.,4,Hangup Regards, Pavel
2019 Apr 23
1
[PATCH v6 5/6] ext4: disable map_sync for async flush
On Tue 23-04-19 13:36:11, Pankaj Gupta wrote: > Dont support 'MAP_SYNC' with non-DAX files and DAX files > with asynchronous dax_device. Virtio pmem provides > asynchronous host page cache flush mechanism. We don't > support 'MAP_SYNC' with virtio pmem and ext4. > > Signed-off-by: Pankaj Gupta <pagupta at redhat.com> The patch looks good to me. You
2019 Jul 17
2
[PATCH v4 4/5] vhost/vsock: split packets to send using multiple buffers
On Wed, Jul 17, 2019 at 01:30:29PM +0200, Stefano Garzarella wrote: > If the packets to sent to the guest are bigger than the buffer > available, we can split them, using multiple buffers and fixing > the length in the packet header. > This is safe since virtio-vsock supports only stream sockets. > > Signed-off-by: Stefano Garzarella <sgarzare at redhat.com> So how does
2019 Jul 17
2
[PATCH v4 4/5] vhost/vsock: split packets to send using multiple buffers
On Wed, Jul 17, 2019 at 01:30:29PM +0200, Stefano Garzarella wrote: > If the packets to sent to the guest are bigger than the buffer > available, we can split them, using multiple buffers and fixing > the length in the packet header. > This is safe since virtio-vsock supports only stream sockets. > > Signed-off-by: Stefano Garzarella <sgarzare at redhat.com> So how does
2019 Apr 23
0
[PATCH v6 5/6] ext4: disable map_sync for async flush
Dont support 'MAP_SYNC' with non-DAX files and DAX files with asynchronous dax_device. Virtio pmem provides asynchronous host page cache flush mechanism. We don't support 'MAP_SYNC' with virtio pmem and ext4. Signed-off-by: Pankaj Gupta <pagupta at redhat.com> --- fs/ext4/file.c | 11 ++++++----- 1 file changed, 6 insertions(+), 5 deletions(-) diff --git
2019 Apr 23
0
[PATCH v6 6/6] xfs: disable map_sync for async flush
Dont support 'MAP_SYNC' with non-DAX files and DAX files with asynchronous dax_device. Virtio pmem provides asynchronous host page cache flush mechanism. We don't support 'MAP_SYNC' with virtio pmem and xfs. Signed-off-by: Pankaj Gupta <pagupta at redhat.com> --- fs/xfs/xfs_file.c | 10 ++++++---- 1 file changed, 6 insertions(+), 4 deletions(-) diff --git
2019 Apr 23
0
[PATCH v6 6/6] xfs: disable map_sync for async flush
On Wed, Apr 24, 2019 at 08:02:17AM +1000, Dave Chinner wrote: > On Tue, Apr 23, 2019 at 01:36:12PM +0530, Pankaj Gupta wrote: > > Dont support 'MAP_SYNC' with non-DAX files and DAX files > > with asynchronous dax_device. Virtio pmem provides > > asynchronous host page cache flush mechanism. We don't > > support 'MAP_SYNC' with virtio pmem and xfs.
2019 Apr 26
0
[PATCH v7 6/6] xfs: disable map_sync for async flush
Dont support 'MAP_SYNC' with non-DAX files and DAX files with asynchronous dax_device. Virtio pmem provides asynchronous host page cache flush mechanism. We don't support 'MAP_SYNC' with virtio pmem and xfs. Signed-off-by: Pankaj Gupta <pagupta at redhat.com> --- fs/xfs/xfs_file.c | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git