similar to: Setting up VHost without QEMU?

Displaying 20 results from an estimated 30000 matches similar to: "Setting up VHost without QEMU?"

2015 Jul 07
0
I can read/write in virtio BLK Device, but I can't run a hello world program in it
On Tue, 7 Jul 2015 10:52:01 +0900 Ganis Zulfa Santoso <ganis.zulfa at gmail.com> wrote: > Hi Linux Virtualization Mailing List Members, > > I am trying to develop a driver for virtio blk device in guest. Why don't you simply use the driver that is already available in the kernel? > I can safely mount the virtual blk device with: mount /dev/vda /mnt/ > in /mnt/ I can
2016 Aug 01
0
[vhost:vhost 14/14] drivers/vhost/vhost.c:915:30: warning: passing argument 2 of 'access_ok' makes pointer from integer without a cast
tree: https://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost.git vhost head: 51ce50c54895044f949129e595ed9a37e4d6c13a commit: 51ce50c54895044f949129e595ed9a37e4d6c13a [14/14] vhost: new device IOTLB API config: sparc64-allmodconfig (attached as .config) compiler: sparc64-linux-gnu-gcc (Debian 5.4.0-6) 5.4.0 20160609 reproduce: wget
2016 Aug 01
0
[vhost:vhost 14/14] drivers/vhost/vhost.c:915:30: warning: passing argument 2 of 'access_ok' makes pointer from integer without a cast
tree: https://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost.git vhost head: 51ce50c54895044f949129e595ed9a37e4d6c13a commit: 51ce50c54895044f949129e595ed9a37e4d6c13a [14/14] vhost: new device IOTLB API config: sparc64-allmodconfig (attached as .config) compiler: sparc64-linux-gnu-gcc (Debian 5.4.0-6) 5.4.0 20160609 reproduce: wget
2015 Nov 17
0
vhost-blk and qemu
Hi,I am looking to experiment the vhost-blk stack. Can some one point me to the latest code version and the corresponding qemu version location.I am on centos 7 (3.10 ) kernel. I am hoping using the vhost-blk.ko and corresponding qemu version will get me started to measure some numbers. RegardsMohan -------------- next part -------------- An HTML attachment was scrubbed... URL:
2015 Nov 17
0
vhost-blk and qemu
Hi,I am looking to experiment the vhost-blk stack. Can some one point me to the latest code version and the corresponding qemu version location.I am on centos 7 (3.10 ) kernel. I am hoping using the vhost-blk.ko and corresponding qemu version will get me started to measure some numbers. RegardsMohan -------------- next part -------------- An HTML attachment was scrubbed... URL:
2014 Nov 06
0
[Qemu-devel] [RFC PATCH 0/2] virtio-mmio: add irqfd support for vhost-net based on virtio-mmio
On 2014/11/5 16:43, Eric Auger wrote: > On 10/27/2014 12:23 PM, Li Liu wrote: >> >> >> On 2014/10/27 17:37, Peter Maydell wrote: >>> On 25 October 2014 09:24, john.liuli <john.liuli at huawei.com> wrote: >>>> To get the interrupt reason to support such VIRTIO_NET_F_STATUS >>>> features I add a new register offset VIRTIO_MMIO_ISRMEM which
2014 Nov 06
0
[Qemu-devel] [RFC PATCH 0/2] virtio-mmio: add irqfd support for vhost-net based on virtio-mmio
On 2014/11/5 16:43, Eric Auger wrote: > On 10/27/2014 12:23 PM, Li Liu wrote: >> >> >> On 2014/10/27 17:37, Peter Maydell wrote: >>> On 25 October 2014 09:24, john.liuli <john.liuli at huawei.com> wrote: >>>> To get the interrupt reason to support such VIRTIO_NET_F_STATUS >>>> features I add a new register offset VIRTIO_MMIO_ISRMEM which
2014 Nov 06
0
[Qemu-devel] [RFC PATCH 0/2] virtio-mmio: add irqfd support for vhost-net based on virtio-mmio
On 2014/11/6 9:59, Shannon Zhao wrote: > > > On 2014/11/5 16:43, Eric Auger wrote: >> On 10/27/2014 12:23 PM, Li Liu wrote: >>> >>> >>> On 2014/10/27 17:37, Peter Maydell wrote: >>>> On 25 October 2014 09:24, john.liuli <john.liuli at huawei.com> wrote: >>>>> To get the interrupt reason to support such VIRTIO_NET_F_STATUS
2015 Sep 01
1
[Qemu-devel] rfc: vhost user enhancements for vm2vm communication
Hi Michael, When you talk about VFIO in guest, is it with a purely emulated IOMMU in Qemu? Also, I am not clear on the following points: 1. How transient memory would be mapped using BAR in the backend VM 2. How would the backend VM update the dirty page bitmap for the frontend VM Regards Varun > -----Original Message----- > From: qemu-devel-bounces+varun.sethi=freescale.com at nongnu.org
2015 Sep 01
1
[Qemu-devel] rfc: vhost user enhancements for vm2vm communication
Hi Michael, When you talk about VFIO in guest, is it with a purely emulated IOMMU in Qemu? Also, I am not clear on the following points: 1. How transient memory would be mapped using BAR in the backend VM 2. How would the backend VM update the dirty page bitmap for the frontend VM Regards Varun > -----Original Message----- > From: qemu-devel-bounces+varun.sethi=freescale.com at nongnu.org
2019 May 23
0
[Qemu-devel] custom virt-io support (in user-mode-linux)
On Wed, May 22, 2019 at 03:02:38PM +0200, Johannes Berg wrote: > Hi, > > While my main interest is mostly in UML right now [1] I've CC'ed the > qemu and virtualization lists because something similar might actually > apply to other types of virtualization. > > I'm thinking about adding virt-io support to UML, but the tricky part is > that while I want to use
2014 Nov 05
2
[Qemu-devel] [RFC PATCH 0/2] virtio-mmio: add irqfd support for vhost-net based on virtio-mmio
On 10/27/2014 12:23 PM, Li Liu wrote: > > > On 2014/10/27 17:37, Peter Maydell wrote: >> On 25 October 2014 09:24, john.liuli <john.liuli at huawei.com> wrote: >>> To get the interrupt reason to support such VIRTIO_NET_F_STATUS >>> features I add a new register offset VIRTIO_MMIO_ISRMEM which >>> will help to establish a shared memory region
2014 Nov 05
2
[Qemu-devel] [RFC PATCH 0/2] virtio-mmio: add irqfd support for vhost-net based on virtio-mmio
On 10/27/2014 12:23 PM, Li Liu wrote: > > > On 2014/10/27 17:37, Peter Maydell wrote: >> On 25 October 2014 09:24, john.liuli <john.liuli at huawei.com> wrote: >>> To get the interrupt reason to support such VIRTIO_NET_F_STATUS >>> features I add a new register offset VIRTIO_MMIO_ISRMEM which >>> will help to establish a shared memory region
2015 Jul 07
2
I can read/write in virtio BLK Device, but I can't run a hello world program in it
Hi Linux Virtualization Mailing List Members, I am trying to develop a driver for virtio blk device in guest. I can safely mount the virtual blk device with: mount /dev/vda /mnt/ in /mnt/ I can read & write the device. But when I run a simple hello world program in /mnt/root/, this error happens: root at linux_guest:/mnt/root# ./helloworld_static [ 23.003459] CPU: 0 PID: 491 Comm:
2015 Jul 07
2
I can read/write in virtio BLK Device, but I can't run a hello world program in it
Hi Linux Virtualization Mailing List Members, I am trying to develop a driver for virtio blk device in guest. I can safely mount the virtual blk device with: mount /dev/vda /mnt/ in /mnt/ I can read & write the device. But when I run a simple hello world program in /mnt/root/, this error happens: root at linux_guest:/mnt/root# ./helloworld_static [ 23.003459] CPU: 0 PID: 491 Comm:
2017 Nov 16
0
[PULL] vhost/virtio/qemu: cleanups and fixes
DMA support in FW CFG had to be pushed out as it caused ltp failures - likely a compatibility issue, and could be a hypervisor bug, but we need to figure it out first. There's still a small chance it'll happen shortly, then I might do another pull request just for that. The following changes since commit bebc6082da0a9f5d47a1ea2edc099bf671058bd4: Linux 4.14 (2017-11-12 10:46:13 -0800)
2017 Nov 16
0
[PULL] vhost/virtio/qemu: cleanups and fixes
DMA support in FW CFG had to be pushed out as it caused ltp failures - likely a compatibility issue, and could be a hypervisor bug, but we need to figure it out first. There's still a small chance it'll happen shortly, then I might do another pull request just for that. The following changes since commit bebc6082da0a9f5d47a1ea2edc099bf671058bd4: Linux 4.14 (2017-11-12 10:46:13 -0800)
2015 Dec 01
1
[RFC PATCH 0/9] vhost-nvme: new qemu nvme backend using nvme target
On 01/12/2015 00:20, Ming Lin wrote: > qemu-nvme: 148MB/s > vhost-nvme + google-ext: 230MB/s > qemu-nvme + google-ext + eventfd: 294MB/s > virtio-scsi: 296MB/s > virtio-blk: 344MB/s > > "vhost-nvme + google-ext" didn't get good enough performance. I'd expect it to be on par of qemu-nvme with ioeventfd but the question is: why should it be better? For
2015 Dec 01
0
[RFC PATCH 0/9] vhost-nvme: new qemu nvme backend using nvme target
On Tue, 2015-12-01 at 17:02 +0100, Paolo Bonzini wrote: > > On 01/12/2015 00:20, Ming Lin wrote: > > qemu-nvme: 148MB/s > > vhost-nvme + google-ext: 230MB/s > > qemu-nvme + google-ext + eventfd: 294MB/s > > virtio-scsi: 296MB/s > > virtio-blk: 344MB/s > > > > "vhost-nvme + google-ext" didn't get good enough performance. > >
2015 Dec 01
1
[RFC PATCH 0/9] vhost-nvme: new qemu nvme backend using nvme target
On 01/12/2015 00:20, Ming Lin wrote: > qemu-nvme: 148MB/s > vhost-nvme + google-ext: 230MB/s > qemu-nvme + google-ext + eventfd: 294MB/s > virtio-scsi: 296MB/s > virtio-blk: 344MB/s > > "vhost-nvme + google-ext" didn't get good enough performance. I'd expect it to be on par of qemu-nvme with ioeventfd but the question is: why should it be better? For