similar to: Xen / NUMA problems

Displaying 20 results from an estimated 4000 matches similar to: "Xen / NUMA problems"

2020 Jul 01
2
no/empty NUMA cells on domain XML
Hi , I'm looking for a possibility to simulate a VM (to be used as a host) reporting no or empty NUMA <cells> in 'virsh capabilities'. 'virsh capabilities' usually reports single NUMA in the VMs, like <numa> <cell id='0' cpus='0-63' memory='16777216' unit='KiB'/> </numa> I appreciate if someone could advise Thank you
2010 Jan 17
2
docs/reference for NUMA usage?
i''m reading up on numa usage. so far, i''ve enabled numa on my Xen box. @ Dom0 i see, xm dmesg | grep -i numa (XEN) Command line: ... numa=on ... (XEN) No NUMA configuration found i guess i need to ''configure'' numa. i''ve no clue how, and haven''t found the docs for it yet, despite looking. i found this old thread,
2019 Sep 23
2
[PATCH RFC v3 1/9] ACPI: NUMA: export pxm_to_node
On 19.09.19 16:22, David Hildenbrand wrote: > Will be needed by virtio-mem to identify the node from a pxm. > > Cc: "Rafael J. Wysocki" <rjw at rjwysocki.net> > Cc: Len Brown <lenb at kernel.org> > Cc: linux-acpi at vger.kernel.org > Signed-off-by: David Hildenbrand <david at redhat.com> > --- > drivers/acpi/numa.c | 1 + > 1 file changed, 1
2006 Oct 04
2
NUMA support on Xen ?
Hi, I am a Masters student from Carnegie Mellon University. I am looking for a research topics for an Advanced OS & DS course we have. I wanted to know what is the current support for NUMA on Xen ? Does it support the IBM x440 and AMD64 Opteron ? Also, does the Xen scheduler do NUMA aware scheduling so it does not degrade the VM performace ? My group is currently looking into Scheduling
2019 Sep 23
2
[PATCH RFC v3 1/9] ACPI: NUMA: export pxm_to_node
On 19.09.19 16:22, David Hildenbrand wrote: > Will be needed by virtio-mem to identify the node from a pxm. > > Cc: "Rafael J. Wysocki" <rjw at rjwysocki.net> > Cc: Len Brown <lenb at kernel.org> > Cc: linux-acpi at vger.kernel.org > Signed-off-by: David Hildenbrand <david at redhat.com> > --- > drivers/acpi/numa.c | 1 + > 1 file changed, 1
2020 Jun 29
2
[RFC 0/3] virtio: NUMA-aware memory allocation
On Sun, Jun 28, 2020 at 02:34:37PM +0800, Jason Wang wrote: > > On 2020/6/25 ??9:57, Stefan Hajnoczi wrote: > > These patches are not ready to be merged because I was unable to measure a > > performance improvement. I'm publishing them so they are archived in case > > someone picks up this work again in the future. > > > > The goal of these patches is to
2020 Jun 29
2
[RFC 0/3] virtio: NUMA-aware memory allocation
On Sun, Jun 28, 2020 at 02:34:37PM +0800, Jason Wang wrote: > > On 2020/6/25 ??9:57, Stefan Hajnoczi wrote: > > These patches are not ready to be merged because I was unable to measure a > > performance improvement. I'm publishing them so they are archived in case > > someone picks up this work again in the future. > > > > The goal of these patches is to
2017 Jun 02
1
should NUMA be enabled?
Hi, should NUMA be enabled in the BIOS of a server that has two sockets but only a single CPU in one of the sockets? From what I?ve been reading, it is unclear to me if NUMA should be enabled only on systems with multiple CPUs in multiple sockets or if multiple cores of a single CPU in a single socket benefit from NUMA being enabled, and if memory access in general benefits from NUMA being
2014 Sep 12
1
Inconsistent behavior between x86_64 and ppc64 when creating guests with NUMA node placement
Hello all, I was recently trying out NUMA placement for my guests on both x86_64 and ppc64 machines. When booting a guest on the x86_64 machine, the following specs were valid (obviously, just notable excepts from the xml): <memory unit='KiB'>8388608</memory> <currentMemory unit='KiB'>8388608</currentMemory> <vcpu
2017 Oct 01
3
NUMA split mode?
Hi, HP says that what they call "NUMA split mode" should be disabled in the BIOS of the Z800 workstation when running Linux. They are reasoning that Linux kernels do not support this feature and even might not boot if it?s enabled. Since it apparently was years ago since they made this statement, I?m wondering if I should still leave this feature disabled or not. More recent kernels
2015 Jan 23
3
questions around using numatune/numa/schedinfo
Hi, I'm running into some problems with libvirt and hoping someone can point me at some instructions or maybe even help me out. First, are there any requirements on qemu version in order to use the "numatune" and/or "cpu/numa/cell" elements? Or do they use cgroups and not the native qemu numa support? Second, are there any instructions on how to set up cgroups? I
2006 Sep 29
4
[PATCH 4/6] xen: export NUMA topology in physinfo hcall
This patch modifies the physinfo hcall to export NUMA CPU and Memory topology information. The new physinfo hcall is integrated into libxc and xend (xm info specifically). Included in this patch is a minor tweak to xm-test''s xm info testcase. The new fields in xm info are: nr_nodes : 4 mem_chunks : node0:0x0000000000000000-0x0000000190000000
2016 Nov 21
2
NUMA VM and assigning interfaces
Hello, Hope all is well. I've been looking at how I can create a virtual machine which is NUMA capable. I was able to do that by: 140 <qemu:commandline> 143 <qemu:arg value='-numa'/> 144 <qemu:arg value='node'/> 145 <qemu:arg value='-numa'/> 146 <qemu:arg value='node'/> 147 </qemu:commandline> This
2017 Oct 02
2
NUMA split mode?
John R Pierce <pierce at hogranch.com> writes: > On 10/1/2017 8:38 AM, hw wrote: >> HP says that what they call "NUMA split mode" should be disabled in the >> BIOS of the Z800 workstation when running Linux. They are reasoning >> that Linux kernels do not support this feature and even might not boot >> if it?s enabled. > > hmm, that workstation is
2005 Dec 16
3
[PATCH] 0/7 xen: Add basic NUMA support
The patchset will add basic NUMA support to Xen (hypervisor only). We borrowed from Linux support for NUMA SRAT table parsing, discontiguous memory tracking (mem chunks), and cpu support (node_to_cpumask etc). The hypervisor parses the SRAT tables and constructs mappings for each node such as node to cpu mappings and memory range to node mappings. Using this information, we also modified the
2018 Aug 17
2
Get Logical processor count correctly whether NUMA is enabled or disabled
Dear R-devel list, R's detectCores() function internally calls "ncpus" function to get the total number of logical processors. However, this doesnot seem to take NUMA into account on Windows machines. On a machine having 48 processors (24 cores) in total and windows server 2012 installed, if NUMA is enabled and has 2 nodes (node 0 and node 1 each having 24 CPUs), then R's
2012 Jul 26
3
[PATCH v8] Some automatic NUMA placement documentation
About rationale, usage and (some small bits of) API. Signed-off-by: Dario Faggioli <dario.faggioli@citrix.com> Acked-by: Ian Campbell <ian.campbell@citrix.com> --- Changes from v7: * avoid referring to 4.2 release as "upcoming". * libxl placement disabling key explicitly mentioned. * Limit of max 16 NUMA nodes explicitly mentioned. Changes from v6: * text updated to
2013 Jul 04
2
Re: [libvirt] [PATCH 1/4] libxl: implement NUMA capabilities reporting
[Moving the conversation on @xen-devel and adding Jan, as that seems more appropriate] [Jan, this came up as I''m implementing some NUMA bits in libvirt but, as you see, the core of Jim''s question is purely about Xen] On lun, 2013-07-01 at 16:47 -0600, Jim Fehlig wrote: > On my non-NUMA test machine I have the cell memory reported as > > <memory
2020 Jun 25
5
[RFC 0/3] virtio: NUMA-aware memory allocation
These patches are not ready to be merged because I was unable to measure a performance improvement. I'm publishing them so they are archived in case someone picks up this work again in the future. The goal of these patches is to allocate virtqueues and driver state from the device's NUMA node for optimal memory access latency. Only guests with a vNUMA topology and virtio devices spread
2020 Jun 25
5
[RFC 0/3] virtio: NUMA-aware memory allocation
These patches are not ready to be merged because I was unable to measure a performance improvement. I'm publishing them so they are archived in case someone picks up this work again in the future. The goal of these patches is to allocate virtqueues and driver state from the device's NUMA node for optimal memory access latency. Only guests with a vNUMA topology and virtio devices spread