search for: zdev

Displaying 20 results from an estimated 51 matches for "zdev".

Did you mean: mdev
2014 Dec 19
2
[PATCH RFC 2/5] s390: add pci_iomap_range
...al mapping cookie for a PCI BAR */ > -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > +void __iomem *pci_iomap_range(struct pci_dev *pdev, > + int bar, > + unsigned long offset, > + unsigned long max) > { > struct zpci_dev *zdev = get_zdev(pdev); > u64 addr; > @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > idx = zdev->bars[bar].map_idx; > spin_lock(&zpci_iomap_lock); > - zpci_iomap_start[idx].fh = zdev->fh; > - zpci_iomap_start[idx]....
2014 Dec 19
2
[PATCH RFC 2/5] s390: add pci_iomap_range
...al mapping cookie for a PCI BAR */ > -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > +void __iomem *pci_iomap_range(struct pci_dev *pdev, > + int bar, > + unsigned long offset, > + unsigned long max) > { > struct zpci_dev *zdev = get_zdev(pdev); > u64 addr; > @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > idx = zdev->bars[bar].map_idx; > spin_lock(&zpci_iomap_lock); > - zpci_iomap_start[idx].fh = zdev->fh; > - zpci_iomap_start[idx]....
2020 Apr 14
0
[PATCH v2 19/33] iommu/s390: Convert to probe/release_device() call-backs
...6,21 +166,14 @@ static void s390_iommu_detach_device(struct iommu_domain *domain, } } -static int s390_iommu_add_device(struct device *dev) +static struct iommu_device *s390_iommu_probe_device(struct device *dev) { - struct iommu_group *group = iommu_group_get_for_dev(dev); struct zpci_dev *zdev = to_pci_dev(dev)->sysdata; - if (IS_ERR(group)) - return PTR_ERR(group); - - iommu_group_put(group); - iommu_device_link(&zdev->iommu_dev, dev); - - return 0; + return &zdev->iommu_dev; } -static void s390_iommu_remove_device(struct device *dev) +static void s390_iommu_relea...
2014 Dec 15
0
[PATCH RFC 2/5] s390: add pci_iomap_range
...ze_t count) } /* Create a virtual mapping cookie for a PCI BAR */ -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) +void __iomem *pci_iomap_range(struct pci_dev *pdev, + int bar, + unsigned long offset, + unsigned long max) { struct zpci_dev *zdev = get_zdev(pdev); u64 addr; @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) idx = zdev->bars[bar].map_idx; spin_lock(&zpci_iomap_lock); - zpci_iomap_start[idx].fh = zdev->fh; - zpci_iomap_start[idx].bar = bar; + if (zpci_iomap_start[...
2014 Dec 15
0
[PATCH RFC 2/5] s390: add pci_iomap_range
...ze_t count) } /* Create a virtual mapping cookie for a PCI BAR */ -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) +void __iomem *pci_iomap_range(struct pci_dev *pdev, + int bar, + unsigned long offset, + unsigned long max) { struct zpci_dev *zdev = get_zdev(pdev); u64 addr; @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) idx = zdev->bars[bar].map_idx; spin_lock(&zpci_iomap_lock); - zpci_iomap_start[idx].fh = zdev->fh; - zpci_iomap_start[idx].bar = bar; + if (zpci_iomap_start[...
2015 Jan 14
0
[PATCH v3 10/16] s390: add pci_iomap_range
...ze_t count) } /* Create a virtual mapping cookie for a PCI BAR */ -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) +void __iomem *pci_iomap_range(struct pci_dev *pdev, + int bar, + unsigned long offset, + unsigned long max) { struct zpci_dev *zdev = get_zdev(pdev); u64 addr; @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) idx = zdev->bars[bar].map_idx; spin_lock(&zpci_iomap_lock); - zpci_iomap_start[idx].fh = zdev->fh; - zpci_iomap_start[idx].bar = bar; + if (zpci_iomap_start[...
2015 Jan 14
0
[PATCH v3 10/16] s390: add pci_iomap_range
...ze_t count) } /* Create a virtual mapping cookie for a PCI BAR */ -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) +void __iomem *pci_iomap_range(struct pci_dev *pdev, + int bar, + unsigned long offset, + unsigned long max) { struct zpci_dev *zdev = get_zdev(pdev); u64 addr; @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) idx = zdev->bars[bar].map_idx; spin_lock(&zpci_iomap_lock); - zpci_iomap_start[idx].fh = zdev->fh; - zpci_iomap_start[idx].bar = bar; + if (zpci_iomap_start[...
2014 Dec 19
0
[PATCH RFC 2/5] s390: add pci_iomap_range
...> > -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > +void __iomem *pci_iomap_range(struct pci_dev *pdev, > > + int bar, > > + unsigned long offset, > > + unsigned long max) > > { > > struct zpci_dev *zdev = get_zdev(pdev); > > u64 addr; > > @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > > > idx = zdev->bars[bar].map_idx; > > spin_lock(&zpci_iomap_lock); > > - zpci_iomap_start[idx].fh = zdev->fh;...
2014 Dec 19
0
[PATCH RFC 2/5] s390: add pci_iomap_range
...> > -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > +void __iomem *pci_iomap_range(struct pci_dev *pdev, > > + int bar, > > + unsigned long offset, > > + unsigned long max) > > { > > struct zpci_dev *zdev = get_zdev(pdev); > > u64 addr; > > @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > > > idx = zdev->bars[bar].map_idx; > > spin_lock(&zpci_iomap_lock); > > - zpci_iomap_start[idx].fh = zdev->fh;...
2008 Nov 26
9
ZPool and Filesystem Sizing - Best Practices?
Hello, We have a new Thor here with 24TB of disk in (first of many, hopefully). We are trying to determine the bext practices with respect to file system management and sizing. Previously, we have tried to keep each file system to a max size of 500GB to make sure we could fit it all on a single tape, and to minimise restore times and impact should we experience some kind of volume
2015 Jan 16
1
[PATCH v3 10/16] s390: add pci_iomap_range
...al mapping cookie for a PCI BAR */ > -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > +void __iomem *pci_iomap_range(struct pci_dev *pdev, > + int bar, > + unsigned long offset, > + unsigned long max) > { > struct zpci_dev *zdev = get_zdev(pdev); > u64 addr; > @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > idx = zdev->bars[bar].map_idx; > spin_lock(&zpci_iomap_lock); > - zpci_iomap_start[idx].fh = zdev->fh; > - zpci_iomap_start[idx]....
2015 Jan 16
1
[PATCH v3 10/16] s390: add pci_iomap_range
...al mapping cookie for a PCI BAR */ > -void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > +void __iomem *pci_iomap_range(struct pci_dev *pdev, > + int bar, > + unsigned long offset, > + unsigned long max) > { > struct zpci_dev *zdev = get_zdev(pdev); > u64 addr; > @@ -270,14 +273,27 @@ void __iomem *pci_iomap(struct pci_dev *pdev, int bar, unsigned long max) > > idx = zdev->bars[bar].map_idx; > spin_lock(&zpci_iomap_lock); > - zpci_iomap_start[idx].fh = zdev->fh; > - zpci_iomap_start[idx]....
2023 Jan 18
10
[PATCH v2 00/10] Let iommufd charge IOPTE allocations to the memory cgroup
iommufd follows the same design as KVM and uses memory cgroups to limit the amount of kernel memory a iommufd file descriptor can pin down. The various internal data structures already use GFP_KERNEL_ACCOUNT to charge its own memory. However, one of the biggest consumers of kernel memory is the IOPTEs stored under the iommu_domain and these allocations are not tracked. This series is the first
2023 Jan 23
11
[PATCH v3 00/10] Let iommufd charge IOPTE allocations to the memory cgroup
iommufd follows the same design as KVM and uses memory cgroups to limit the amount of kernel memory a iommufd file descriptor can pin down. The various internal data structures already use GFP_KERNEL_ACCOUNT to charge its own memory. However, one of the biggest consumers of kernel memory is the IOPTEs stored under the iommu_domain and these allocations are not tracked. This series is the first
2023 Jan 23
11
[PATCH v3 00/10] Let iommufd charge IOPTE allocations to the memory cgroup
iommufd follows the same design as KVM and uses memory cgroups to limit the amount of kernel memory a iommufd file descriptor can pin down. The various internal data structures already use GFP_KERNEL_ACCOUNT to charge its own memory. However, one of the biggest consumers of kernel memory is the IOPTEs stored under the iommu_domain and these allocations are not tracked. This series is the first
2023 Jan 06
8
[PATCH 0/8] Let iommufd charge IOPTE allocations to the memory cgroup
iommufd follows the same design as KVM and uses memory cgroups to limit the amount of kernel memory a iommufd file descriptor can pin down. The various internal data structures already use GFP_KERNEL_ACCOUNT to charge its own memory. However, one of the biggest consumers of kernel memory is the IOPTEs stored under the iommu_domain and these allocations are not tracked. This series is the first
2023 Jan 06
8
[PATCH 0/8] Let iommufd charge IOPTE allocations to the memory cgroup
iommufd follows the same design as KVM and uses memory cgroups to limit the amount of kernel memory a iommufd file descriptor can pin down. The various internal data structures already use GFP_KERNEL_ACCOUNT to charge its own memory. However, one of the biggest consumers of kernel memory is the IOPTEs stored under the iommu_domain and these allocations are not tracked. This series is the first
2023 Jan 06
8
[PATCH 0/8] Let iommufd charge IOPTE allocations to the memory cgroup
iommufd follows the same design as KVM and uses memory cgroups to limit the amount of kernel memory a iommufd file descriptor can pin down. The various internal data structures already use GFP_KERNEL_ACCOUNT to charge its own memory. However, one of the biggest consumers of kernel memory is the IOPTEs stored under the iommu_domain and these allocations are not tracked. This series is the first
2014 Dec 15
6
[PATCH RFC 0/5] virtio pci: virtio 1.0 support
This is on top of 3.19 master + my bugfix patches, and adds virtio 1.0 support to virtio pci. This is 3.20 material I think. Would like to get feedback on s390 change as it's untested. Michael S Tsirkin (2): pci: add pci_iomap_range s390: add pci_iomap_range Michael S. Tsirkin (2): virtio_pci: modern driver virtio_pci: macros for PCI layout offsets. Rusty Russell (1): virtio-pci:
2014 Dec 15
6
[PATCH RFC 0/5] virtio pci: virtio 1.0 support
This is on top of 3.19 master + my bugfix patches, and adds virtio 1.0 support to virtio pci. This is 3.20 material I think. Would like to get feedback on s390 change as it's untested. Michael S Tsirkin (2): pci: add pci_iomap_range s390: add pci_iomap_range Michael S. Tsirkin (2): virtio_pci: modern driver virtio_pci: macros for PCI layout offsets. Rusty Russell (1): virtio-pci: