similar to: docs/reference for NUMA usage?

Displaying 20 results from an estimated 5000 matches similar to: "docs/reference for NUMA usage?"

2010 Sep 19
4
Upgrade Xen up to 4.0.1 on openSUSE 11.3 and broken VFB
After upgrade via http://download.opensuse.org/repositories/Virtualization:/openSUSE11.3/openSUSE_11.3/src/xen-4.0.1_21326_01-1.1.src.rpm with any "kotd" or 2.6.34.4-2.2-xen VFB gets broken . It worked with original  Xen 4.0, no matter of  some additional patching to support ZFS 24 has been done and 4.0 src.rpm has been rebuilt to reinstall 4.0   In particular,  virt-install fails to
2010 Sep 08
5
domU paravirt vfb problem
Hello! I''m investigate strange behaviour of vfb in paravirt domU (centos 5) VNC servers started after creating domain, but information in xenstore appears sometimes after 1-2 minutes after start. Why this happening? And how can i get VNC port immediate after creating of domU? Thank''s. -- Vasiliy G Tolstov <v.tolstov@selfip.ru> Selfip.Ru
2008 May 21
2
outgoing domu network dies after a while
I have a OpenSuse 10.3 with xen 3.1.0 running and it''s been running fine for a few months. This past weekend it suddenly started to act up and after some troubleshooting I can now say that it seems like the guests(domU) loose the outgoing network pipe, from the console I can see that the TX counter is stuck at the same value but it''s no errors. It behaves as if whatever I try to
2013 Sep 04
10
Performance test regarding xenstored
Hello, I am running a mail benchmark test using smtp-source[1] with VM running postfix on Xen hypervisor. My system configuration is: uname -a -> Linux cadlab 3.1.10-1.19-xen #1 SMP Mon Feb 25 10:32:50 UTC 2013 (f0b13a3) x86_64 x86_64 x86_64 GNU/Linux I am observing a very high disk write usage by xenstored (some 5 Mbps) without even anything running on the VM. Is it normal? During test, I
2013 Sep 04
10
[Xen-users] Performance test regarding xenstored
Hello, I am running a mail benchmark test using smtp-source[1] with VM running postfix on Xen hypervisor. My system configuration is: uname -a -> Linux cadlab 3.1.10-1.19-xen #1 SMP Mon Feb 25 10:32:50 UTC 2013 (f0b13a3) x86_64 x86_64 x86_64 GNU/Linux I am observing a very high disk write usage by xenstored (some 5 Mbps) without even anything running on the VM. Is it normal? During test, I
2011 Mar 26
2
bug in xen 4.1 install on suse 11.4 (xencommons)
List, here''s how I got 4.1 up on suse 11.4 There is a small bug with the xencommons init script preventing xenstored from starting. Add this repo for 4.1 RPM''s ------------- zypper ar http://download.opensuse.org/repositories/Virtualization/openSUSE_11.4 Virt zypper ref zypper in xen-4.1.0_01-106.1 Fix the xen inits  (observe the order here) ------------ chkconfig -s
2006 Oct 24
11
Xen 3.0.3 confusion
I am thoughly confused with the various Xen 3.0.3 versions for Suse. Firt there is not for suse 10.1, so I gues I woulllt the source tarball and compile. ( I tried the RPM for 10.0 (I am running SuSE 10.1) but it caused a kernel panic) I have 5 packages listed unter Yast software management (not including documentation,pdf, etc). Kernel-xen ver 2.6.16.21.25 Xen 3.02_09763-0.8 xen-libs
2008 Jul 31
1
Am I running NUMA?
The subject says it all really, I''m pretty sure the system and bios are NUMA aware but I''m not sure if it''s enabled? Is it supported in 3.2.1 stable? Thanks, Rob \ \/ /___ _ __ |___ / |___ \ / | \ // _ \ ''_ \ |_ \ __) | | | / \ __/ | | | ___) | / __/ _| | /_/\_\___|_| |_| |____(_)_____(_)_| (XEN) Xen version 3.2.1
2018 Sep 14
3
NUMA issues on virtualized hosts
Hello, I have cluster with AMD EPYC 7351 cpu. Two CPUs per node. I have performance 8-NUMA configuration: This is from hypervizor: [root@hde10 ~]# lscpu Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 64 On-line CPU(s) list: 0-63 Thread(s) per core: 2 Core(s) per socket: 16 Socket(s): 2 NUMA
2018 Sep 14
1
Re: NUMA issues on virtualized hosts
Hello again, when the iozone writes slow. This is how slabtop looks like: 62476752 62476728 0% 0.10K 1601968 39 6407872K buffer_head 1000678 999168 0% 0.56K 142954 7 571816K radix_tree_node 132184 125911 0% 0.03K 1066 124 4264K kmalloc-32 118496 118224 0% 0.12K 3703 32 14812K kmalloc-node 73206 56467 0% 0.19K 3486 21
2010 Dec 30
2
Having stability problems with Linux-2.6.34.7-0.5-xen
Hello, I am using Xen 4.0.0_21091_06-0.1 on OpenSuse 11.3 64 bit, with a Windows XP Pro SP3 32 bit guest, and my workstation keeping locking up, even the NumLock and CapsLock keys being unresponsive when it does. Where should I begin troubleshooting? Thanks, Todd CPU info: ___________________________________________________________________________ processor : 0 vendor_id : GenuineIntel
2006 Sep 29
4
[PATCH 4/6] xen: export NUMA topology in physinfo hcall
This patch modifies the physinfo hcall to export NUMA CPU and Memory topology information. The new physinfo hcall is integrated into libxc and xend (xm info specifically). Included in this patch is a minor tweak to xm-test''s xm info testcase. The new fields in xm info are: nr_nodes : 4 mem_chunks : node0:0x0000000000000000-0x0000000190000000
2008 Sep 05
0
3.2.1+ HVM + HAP + NUMA - Poor Memory Performance
Hi Everyone, I am running 3.2.1 on Centos 5.2 with HAP enabled, NUMA enabled, ACPI enabled and the dom0 allocated 512Mb. I have setup a single core 1Gb VM for performance testing under Windows 2008 Server. Most CPU results are within a few percent of theoretical max but Memory performance is about half what I expected. I get 3.22Gb/Sec Sandra 2009 Memory performance for a single Opteron 8350
2020 Jul 01
2
no/empty NUMA cells on domain XML
Hi , I'm looking for a possibility to simulate a VM (to be used as a host) reporting no or empty NUMA <cells> in 'virsh capabilities'. 'virsh capabilities' usually reports single NUMA in the VMs, like <numa> <cell id='0' cpus='0-63' memory='16777216' unit='KiB'/> </numa> I appreciate if someone could advise Thank you
2020 Jun 29
2
[RFC 0/3] virtio: NUMA-aware memory allocation
On Sun, Jun 28, 2020 at 02:34:37PM +0800, Jason Wang wrote: > > On 2020/6/25 ??9:57, Stefan Hajnoczi wrote: > > These patches are not ready to be merged because I was unable to measure a > > performance improvement. I'm publishing them so they are archived in case > > someone picks up this work again in the future. > > > > The goal of these patches is to
2020 Jun 29
2
[RFC 0/3] virtio: NUMA-aware memory allocation
On Sun, Jun 28, 2020 at 02:34:37PM +0800, Jason Wang wrote: > > On 2020/6/25 ??9:57, Stefan Hajnoczi wrote: > > These patches are not ready to be merged because I was unable to measure a > > performance improvement. I'm publishing them so they are archived in case > > someone picks up this work again in the future. > > > > The goal of these patches is to
2014 Sep 12
1
Inconsistent behavior between x86_64 and ppc64 when creating guests with NUMA node placement
Hello all, I was recently trying out NUMA placement for my guests on both x86_64 and ppc64 machines. When booting a guest on the x86_64 machine, the following specs were valid (obviously, just notable excepts from the xml): <memory unit='KiB'>8388608</memory> <currentMemory unit='KiB'>8388608</currentMemory> <vcpu
2019 Sep 23
2
[PATCH RFC v3 1/9] ACPI: NUMA: export pxm_to_node
On 19.09.19 16:22, David Hildenbrand wrote: > Will be needed by virtio-mem to identify the node from a pxm. > > Cc: "Rafael J. Wysocki" <rjw at rjwysocki.net> > Cc: Len Brown <lenb at kernel.org> > Cc: linux-acpi at vger.kernel.org > Signed-off-by: David Hildenbrand <david at redhat.com> > --- > drivers/acpi/numa.c | 1 + > 1 file changed, 1
2019 Sep 23
2
[PATCH RFC v3 1/9] ACPI: NUMA: export pxm_to_node
On 19.09.19 16:22, David Hildenbrand wrote: > Will be needed by virtio-mem to identify the node from a pxm. > > Cc: "Rafael J. Wysocki" <rjw at rjwysocki.net> > Cc: Len Brown <lenb at kernel.org> > Cc: linux-acpi at vger.kernel.org > Signed-off-by: David Hildenbrand <david at redhat.com> > --- > drivers/acpi/numa.c | 1 + > 1 file changed, 1
2017 Jun 02
1
should NUMA be enabled?
Hi, should NUMA be enabled in the BIOS of a server that has two sockets but only a single CPU in one of the sockets? From what I?ve been reading, it is unclear to me if NUMA should be enabled only on systems with multiple CPUs in multiple sockets or if multiple cores of a single CPU in a single socket benefit from NUMA being enabled, and if memory access in general benefits from NUMA being