search for: dma_devic

Displaying 20 results from an estimated 114 matches for "dma_devic".

Did you mean: dma_device
2020 Mar 04
1
[PATCH V5 3/5] vDPA: introduce vDPA bus
On Wed, Feb 26, 2020 at 02:04:54PM +0800, Jason Wang wrote: > +struct vdpa_device *vdpa_alloc_device(struct device *parent, > + struct device *dma_dev, > + const struct vdpa_config_ops *config) > +{ > + struct vdpa_device *vdev; > + int err = -ENOMEM; > + > + if (!parent || !dma_dev || !config) > + goto err; > + > + vdev = kzalloc(sizeof(*vdev),
2019 May 23
0
[PATCH v2 2/8] s390/cio: introduce DMA pools to cio
From: Halil Pasic <pasic at linux.ibm.com> To support protected virtualization cio will need to make sure the memory used for communication with the hypervisor is DMA memory. Let us introduce one global cio, and some tools for pools seated at individual devices. Our DMA pools are implemented as a gen_pool backed with DMA pages. The idea is to avoid each allocation effectively wasting a
2019 Apr 26
0
[PATCH 05/10] s390/cio: introduce DMA pools to cio
To support protected virtualization cio will need to make sure the memory used for communication with the hypervisor is DMA memory. Let us introduce one global cio, and some tools for pools seated at individual devices. Our DMA pools are implemented as a gen_pool backed with DMA pages. The idea is to avoid each allocation effectively wasting a page, as we typically allocate much less than
2023 Apr 02
2
[PATCH] virtio-vdpa: add VIRTIO_F_NOTIFICATION_DATA feature support
Add VIRTIO_F_NOTIFICATION_DATA support for vDPA transport. If this feature is negotiated, the driver passes extra data when kicking a virtqueue. A device that offers this feature needs to implement the kick_vq_with_data callback. kick_vq_with_data receives the vDPA device and data. data includes the vqn, next_off and next_wrap for packed virtqueues. This patch follows a patch [1] by Viktor
2019 May 29
0
[PATCH v3 2/8] s390/cio: introduce DMA pools to cio
From: Halil Pasic <pasic at linux.ibm.com> To support protected virtualization cio will need to make sure the memory used for communication with the hypervisor is DMA memory. Let us introduce one global pool for cio. Our DMA pools are implemented as a gen_pool backed with DMA pages. The idea is to avoid each allocation effectively wasting a page, as we typically allocate much less than
2019 May 27
3
[PATCH v2 2/8] s390/cio: introduce DMA pools to cio
On Thu, 23 May 2019 18:22:03 +0200 Michael Mueller <mimu at linux.ibm.com> wrote: > From: Halil Pasic <pasic at linux.ibm.com> > > To support protected virtualization cio will need to make sure the > memory used for communication with the hypervisor is DMA memory. > > Let us introduce one global cio, and some tools for pools seated "one global pool for
2019 May 27
3
[PATCH v2 2/8] s390/cio: introduce DMA pools to cio
On Thu, 23 May 2019 18:22:03 +0200 Michael Mueller <mimu at linux.ibm.com> wrote: > From: Halil Pasic <pasic at linux.ibm.com> > > To support protected virtualization cio will need to make sure the > memory used for communication with the hypervisor is DMA memory. > > Let us introduce one global cio, and some tools for pools seated "one global pool for
2019 Jun 06
0
[PATCH v4 2/8] s390/cio: introduce DMA pools to cio
To support protected virtualization cio will need to make sure the memory used for communication with the hypervisor is DMA memory. Let us introduce one global pool for cio. Our DMA pools are implemented as a gen_pool backed with DMA pages. The idea is to avoid each allocation effectively wasting a page, as we typically allocate much less than PAGE_SIZE. Signed-off-by: Halil Pasic <pasic at
2019 Jun 12
0
[PATCH v5 2/8] s390/cio: introduce DMA pools to cio
To support protected virtualization cio will need to make sure the memory used for communication with the hypervisor is DMA memory. Let us introduce one global pool for cio. Our DMA pools are implemented as a gen_pool backed with DMA pages. The idea is to avoid each allocation effectively wasting a page, as we typically allocate much less than PAGE_SIZE. Signed-off-by: Halil Pasic <pasic at
2023 Feb 21
2
[PATCH vhost 07/10] virtio_ring: add api virtio_dma_map() for advance dma
On Mon, Feb 20, 2023 at 3:02 PM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > On Mon, 20 Feb 2023 13:38:24 +0800, Jason Wang <jasowang at redhat.com> wrote: > > On Tue, Feb 14, 2023 at 3:27 PM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > > > > Added virtio_dma_map() to map DMA addresses for virtual memory in > > > advance.
2023 Feb 21
2
[PATCH vhost 07/10] virtio_ring: add api virtio_dma_map() for advance dma
On Mon, Feb 20, 2023 at 3:02 PM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > On Mon, 20 Feb 2023 13:38:24 +0800, Jason Wang <jasowang at redhat.com> wrote: > > On Tue, Feb 14, 2023 at 3:27 PM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > > > > Added virtio_dma_map() to map DMA addresses for virtual memory in > > > advance.
2023 Feb 16
0
[PATCH v2] vhost/vdpa: Add MSI translation tables to iommu for software-managed MSI
? 2023/2/7 20:08, Nanyong Sun ??: > From: Rong Wang <wangrong68 at huawei.com> > > Once enable iommu domain for one device, the MSI > translation tables have to be there for software-managed MSI. > Otherwise, platform with software-managed MSI without an > irq bypass function, can not get a correct memory write event > from pcie, will not get irqs. > The solution is
2020 Aug 05
2
[PATCH v2 19/24] vdpa: make sure set_features in invoked for legacy
On 2020/8/4 ??5:00, Michael S. Tsirkin wrote: > Some legacy guests just assume features are 0 after reset. > We detect that config space is accessed before features are > set and set features to 0 automatically. > Note: some legacy guests might not even access config space, if this is > reported in the field we might need to catch a kick to handle these. I wonder whether it's
2020 Aug 05
2
[PATCH v2 19/24] vdpa: make sure set_features in invoked for legacy
On 2020/8/4 ??5:00, Michael S. Tsirkin wrote: > Some legacy guests just assume features are 0 after reset. > We detect that config space is accessed before features are > set and set features to 0 automatically. > Note: some legacy guests might not even access config space, if this is > reported in the field we might need to catch a kick to handle these. I wonder whether it's
2023 Mar 02
1
[PATCH vhost 07/10] virtio_ring: add api virtio_dma_map() for advance dma
On Tue, 28 Feb 2023 19:15:23 +0800, Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > On Tue, 21 Feb 2023 09:51:07 +0800, Jason Wang <jasowang at redhat.com> wrote: > > On Mon, Feb 20, 2023 at 3:02 PM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > > > > On Mon, 20 Feb 2023 13:38:24 +0800, Jason Wang <jasowang at redhat.com> wrote: >
2020 Feb 10
0
[PATCH V2 3/5] vDPA: introduce vDPA bus
vDPA device is a device that uses a datapath which complies with the virtio specifications with vendor specific control path. vDPA devices can be both physically located on the hardware or emulated by software. vDPA hardware devices are usually implemented through PCIE with the following types: - PF (Physical Function) - A single Physical Function - VF (Virtual Function) - Device that supports
2023 Mar 02
1
[PATCH vhost 07/10] virtio_ring: add api virtio_dma_map() for advance dma
On Thu, Mar 2, 2023 at 10:24?AM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > On Tue, 28 Feb 2023 19:15:23 +0800, Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > On Tue, 21 Feb 2023 09:51:07 +0800, Jason Wang <jasowang at redhat.com> wrote: > > > On Mon, Feb 20, 2023 at 3:02 PM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > >
2023 Mar 02
2
[PATCH vhost 07/10] virtio_ring: add api virtio_dma_map() for advance dma
On Thu, Mar 02, 2023 at 11:26:53AM +0800, Jason Wang wrote: > On Thu, Mar 2, 2023 at 11:24?AM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > > On Thu, 2 Mar 2023 11:05:26 +0800, Jason Wang <jasowang at redhat.com> wrote: > > > On Thu, Mar 2, 2023 at 10:24?AM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > > > > > > On
2020 Feb 20
0
[PATCH V3 3/5] vDPA: introduce vDPA bus
vDPA device is a device that uses a datapath which complies with the virtio specifications with vendor specific control path. vDPA devices can be both physically located on the hardware or emulated by software. vDPA hardware devices are usually implemented through PCIE with the following types: - PF (Physical Function) - A single Physical Function - VF (Virtual Function) - Device that supports
2023 Mar 02
1
[PATCH vhost 07/10] virtio_ring: add api virtio_dma_map() for advance dma
On Thu, 2 Mar 2023 11:05:26 +0800, Jason Wang <jasowang at redhat.com> wrote: > On Thu, Mar 2, 2023 at 10:24?AM Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > > On Tue, 28 Feb 2023 19:15:23 +0800, Xuan Zhuo <xuanzhuo at linux.alibaba.com> wrote: > > > On Tue, 21 Feb 2023 09:51:07 +0800, Jason Wang <jasowang at redhat.com> wrote: > >