similar to: [PATCH 0/5] Add MSI support to XEN

Displaying 20 results from an estimated 8000 matches similar to: "[PATCH 0/5] Add MSI support to XEN"

2008 Apr 10
8
[PATCH][RFC]Move PCI Configuration Spaces from Dom0 to Xen
Hi, Keir, This patch will move reading and writing of PCI configuration spaces from dom0 to Xen. It also changes VTD code, so that they can touch the PCI configuration spaces with proper lock. This will also benefit MSI support in Xen. Can you give some comments? Thanks! <<pci_conf_xen.patch>> Best Regards Haitao Shan _______________________________________________ Xen-devel
2008 Mar 27
11
[PATCH 1/5] Add MSI support to XEN
This patch changes the pirq to be per-domain in xen tree. Signed-off-by: Jiang Yunhong <yunhong.jiang@intel.com> Signed-off-by: Shan Haitao <haitao.shan@intel.com> Best Regards Shan Haitao _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel
2008 Jul 03
13
[PATCH] Handle MSI irq storm
<<handle_msi_irq_storm.patch>> Hi, Keir, This patch handles MSI irq storm. Unluckily, I have observed this phenomenon again. This will happen when some kind of MSI-X capable NIC is assigned to an HVM guest. The basic idea is to mask the interrupt on receiving the second interrupt and set a timer to unmask after 1ms. Can you have a look and give some comments on that? Thanks! Best
2008 Sep 09
29
[PATCH 1/4] CPU online/offline support in Xen
This patch implements cpu offline feature. Best Regards Haitao Shan _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com http://lists.xensource.com/xen-devel
2008 Feb 20
5
MSI Support in Xen
I was trying to patch Xen 3.2 from the mercurial repository with the MSI patches by Yunhong Jiang given at the bottom of this page : http://lists.xensource.com/archives/html/xen-devel/2007-11/msg00157.html The tree seems to have significantly changed, especially the hvm_irq_dpci_mapping structure in irq.h header file. Are there any updated patches, or can I get a version of Xen from somewhere
2008 Jul 01
4
VMX status report. Xen: #17917 & Xen0: #583 -- no new issue
Hi all, This is our nightly testing report for c/s 17917; no new issue today, and bug #1286 got fixed. Fixed issue: ============================================== 1. guest cannot be created with two disks appended. http://bugzilla.xensource.com/bugzilla/show_bug.cgi?id=1286 Old issues: ============================================== 1. dom0 hangs with io errors printed while booting hvm
2007 Dec 11
13
[PATCH] Enable Core 2 Duo Performance Counters in HVM guest
Hi, Keir, Currently, HVM guests do not have access to performance counters. So it is not possible to use performance analyzer software such as vtune in HVM guest to analyze programme performance. Other usage of performance counters , for example, the NMI watchdog, won''t function either. This patch will enable performance counters in HVM guest. Currently, only Core 2 Duo is implemented.
2008 Nov 18
6
[PATCH] fix memory allocation from NUMA node for VT-d.
The memory relating guest domain should be allocated from NUMA node on which the guest runs. Because the latency of the same NUMA node is faster than that of a different one. This patch fixes memory allocation for Address Translation Structure of VT-d. VT-d uses two types of Structures for DMA address translation. The one is Device Assignment Structure. The other is Address Translation
2008 Dec 08
4
[PATCH][VTD] pci mmcfg patch for x86-64 - version 2
Fixes made in version 2: 1) Use PML4[257] for ioremap of PCI mmcfg. As full 16-bit segment support would require 44-bits. Since each slot only has 39-bits, we support 2048 PCI segments for now. This can be easily expanded if deemed necessary in the future. 2) Integrated PCI mmcfg access with existing PCI config interface for x86_64. Use MMCFG interface if offset is greater than 256.
2008 May 19
21
[PATCH 0/5] VT-d support for PV guests
Hi, I''ve added some preliminary support for VT-d for paravirtualized guests. This must be enabled using an ''iommu_pv'' boot parameter (disabled by default). I''ve added some python bindigs to allow xend to assign PCI devices to IOMMU for PV guests. For HVM guests this is handled in ioemu. Not sure if it makes sense to handle both cases in one place. The
2008 May 27
3
[PATCH] VT-d: IOTLB flush fixups
On map: only flush when old PTE was valid or invalid PTE may be cached. On unmap: always flush old entry, but skip flush for unaffected IOMMUs. Signed-off-by: Espen Skoglund <espen.skoglund@netronome.com> -- iommu.c | 17 +++++++++++------ 1 file changed, 11 insertions(+), 6 deletions(-) -- diff -r 8187fd8113f9 xen/drivers/passthrough/vtd/iommu.c ---
2008 Sep 11
12
pciback module not working
This issue was brought up about a year and a half ago. So I''ll bring it up again and see if anything happens. I''ve got an infiniband network and am attempting to pass the infiniband card through the host and give it to the guest. I''m working with standard CentOS 5.2 on both guest and host with their provided xen (3.0.3 ish). I''ve also attempted to install the
2008 Sep 11
12
pciback module not working
This issue was brought up about a year and a half ago. So I''ll bring it up again and see if anything happens. I''ve got an infiniband network and am attempting to pass the infiniband card through the host and give it to the guest. I''m working with standard CentOS 5.2 on both guest and host with their provided xen (3.0.3 ish). I''ve also attempted to install the
2008 Sep 17
7
Megaraid SAS driver failing in Xen-3.3.0 but was working in Xen-3.2.2-rc3
On Xen-3.3.0, domain0 Megaraid SAS (SAS 1068 controller) driver is not loading correctly if vtd support in Xen is enabled. It fails at the point of initializing firmware. I wasn''t seeing this error with Xen-3.2.2-rc3 (Unstable version), though with vtd disabled in Xen-3.3.0, it is working. Looks like a degrade problem. Any clues? Thx, Venkat
2008 Sep 19
19
MSI causing softpanics in guest
lspci shows MSI enabled for PCI device. PCI passthrough works fine. However, as soon as the MSI driver for card is insmodded, kernel panics. This is on xen-unstable. Tried the same with xen-3.3.0 which is supposed to have MSI passthrough, but the same guest shows MSI as disabled. Any else seen this bug, or know of a workaround ? Trace is as follows : ------------[ cut here ]------------
2008 Jul 12
26
[PATCH] Improve the current FLR logic
Hi, all, The attached patches try to improve the current FLR logic. The idea is: removing the FLR logic from hypervisor and adding the improved logic in Control Panel. The current FLR logic in hypervisor has some issues: 1) Dstate transition is not guaranteed to properly clear the device state; 2) the current code for PCIe FLR is actually buggy: PCI_EXP_DEVSTA_TRPND doesn''t mean the
2007 Jun 28
5
One question on MMIO
In sh_page_fault(), there are some code like following, why we think it is mmio only when paging_vcpu_mode_translate(v)? Thanks Yunhong Jiang /* What mfn is the guest trying to access? */ gfn = guest_l1e_get_gfn(gw.eff_l1e); gmfn = vcpu_gfn_to_mfn(v, gfn); mmio = (is_hvm_domain(d) && paging_vcpu_mode_translate(v) &&
2008 Sep 23
9
Xen crash on dom0 shutdown
There is a BUG_ON() at xen/arch/x86/physdev.c:169 which appears to be dependent upon guest behavior (should close event channel before un-mapping pirq), rather than on internal hypervisor state. In 2.6.18, this likely goes unnoticed because pci_device_shutdown() only calls all the driver shutdown routines. In newer kernels, however, it also calls pci_msi_shutdown() and pci_msix_shutdown(), which
2011 Jul 12
4
Possible bug? Can't remove and then put back a file?
This happens to Puppet 2.7.1 (both master and client). I have a file resource that I set to ensure => absent, Puppet removed it correctly with info line that it has a duplicate entry in the filebucket. I then changed the ensure => present, the agent run gave me following error: /usr/bin/puppet:19 err: /Stage[main]/Emb/File[/opt/jetty-distribution-7.4.2.v20110526.tar.gz]/ensure: change from
2008 Dec 15
8
[PATCH 0/2] MCA support with page offlining
Hi all, I had posted about MCA support for Intel64 before. It had only a function to log the MCA error data received from hypervisor. http://lists.xensource.com/archives/html/xen-devel/2008-09/msg00876.html I attach patches that support not only error logging but also Page Offlining function. The page where an MCA occurs will offline and not reuse. A new flag ''PGC_reserved''