similar to: help about migration with libvirt

Displaying 20 results from an estimated 2000 matches similar to: "help about migration with libvirt"

2013 Dec 03
1
Re: help about migration with libvirt
On 12/03/2013 12:52 AM, Nicolas Sebrecht wrote: > This is about CPU affinity. You may look at CPU pinning with 'vcpupin' > in the manual of virsh. It is possible to enable vCPU pinning in libvirt > guest XML. > > Don't know if libvirt supports applying CPU affinity at guest runtime Yes, 'virsh vcpupin' can be used to change CPU affinity at runtime. > but
2013 Mar 12
2
managedsave results in unexpected shutdown from inside Windows
Hi, I'm facing a strange behaviour of windows 2008 guests. After a virsh managedsave and start, the windows guest open a "unexpected shutdown" window. The window looks like this one (randomly taken from the web): http://toastytech.com/guis/srv2k3login2.jpg This happens in all virtualized windows in a hypervisor. Has anyone else experienced such issue? Thanks, -- Nicolas
2018 Sep 05
2
Domain vCPU threads affinity
Hello, According to the docs, vcpupin will use either cgroups or sched_setaffinity to pin vcpu threads to cpus. How is this decision made? I observe differences even on different hosts featuring the same version of libvirtd (1.3.1): on one host vcpupin affects cpuset.cpus (cgroup), and on the other it affects vcpu threads affinity (observed through taskset). Thanks, Nikos -------------- next
2012 Apr 04
1
vcpupin and kvm
Hi. Can anyone confirm if vcpupin works correctly with kvm? If set I pin a vcpu it shows up correctly via vcpuinfo, but querying via taskset shows the cpu is not pinned. Thanks, -Geoff -- ---- Geoff Galitz, ggalitz at shutterstock.com WebOps Engineer, Europe Shutterstock Images http://wwww.shutterstock.com/
2010 May 21
10
What''s the different for "dom0_max_vcpus=4 dom0_vcpus_pin" and "dom0_max_vcpus=4" ?
Hi experts, Q1:What''s the different for "dom0_max_vcpus=4 dom0_vcpus_pin" and "dom0_max_vcpus=4" ? which will get better performance Q2: dom0_max_vcpus=4 means "core0-3 will be just used by dom0" or means "4 cores(not dedicate cores) will be used by dom0, eg: core2-5 or core3-6? Q3.what does mean "nosmp" , xen, dom0,domU, will just use one
2010 May 21
10
What''s the different for "dom0_max_vcpus=4 dom0_vcpus_pin" and "dom0_max_vcpus=4" ?
Hi experts, Q1:What''s the different for "dom0_max_vcpus=4 dom0_vcpus_pin" and "dom0_max_vcpus=4" ? which will get better performance Q2: dom0_max_vcpus=4 means "core0-3 will be just used by dom0" or means "4 cores(not dedicate cores) will be used by dom0, eg: core2-5 or core3-6? Q3.what does mean "nosmp" , xen, dom0,domU, will just use one
2011 May 11
1
bizzare performance problem
I had a rather strange problem last week with one of our 8 core servers. The users complained the performance was "slow" so I checked the basic things, processes on top, vmstat for memory and context switching, i/o stats for internal disk I/O, netstat for any network issues and other things like network through put by copying a large file (1gb file across the network). It turned out I
2013 Sep 06
21
[PATCH v2 0/5] xl: allow for node-wise specification of vcpu pinning
Hi all, This is the second take of a patch that I submitted some time ago for allowing specifying vcpu pinning taking NUMA nodes into account. IOW, something like this: * "nodes:0-3": all pCPUs of nodes 0,1,2,3;  * "nodes:0-3,^node:2": all pCPUS of nodes 0,1,3;  * "1,nodes:1-2,^6": pCPU 1 plus all pCPUs of nodes 1,2    but not pCPU 6; v1 was a single patch, this is
2013 Dec 03
0
Re: help about migration with libvirt
The 03/12/13, lei yang wrote: > Hi experts > > I want to use libvirt doing the migration Basically, migration allows moving a VM from host to another. > migration the guest from core 0 to core 2, Is there a way to do this ? any > points? This is about CPU affinity. You may look at CPU pinning with 'vcpupin' in the manual of virsh. It is possible to enable
2018 Sep 17
2
Re: NUMA issues on virtualized hosts
On 09/14/2018 03:36 PM, Lukas Hejtmanek wrote: > Hello, > > ok, I found that cpu pinning was wrong, so I corrected it to be 1:1. The issue > with iozone remains the same. > > The spec is running, however, it runs slower than 1-NUMA case. > > The corrected XML looks like follows: [Reformated XML for better reading] <cpu mode="host-passthrough">
2018 Sep 14
3
NUMA issues on virtualized hosts
Hello, I have cluster with AMD EPYC 7351 cpu. Two CPUs per node. I have performance 8-NUMA configuration: This is from hypervizor: [root@hde10 ~]# lscpu Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 64 On-line CPU(s) list: 0-63 Thread(s) per core: 2 Core(s) per socket: 16 Socket(s): 2 NUMA
2010 May 19
4
xen4.0.0 64 bit boot "panic on cpu 0, crc errror"
Hi experts, I boot the xen4.0.0 with gpxe and boot xen 32 bit works well but when I try boot xen 64 bit, the error happens (XEN)Brought up 16 cpus (XEN) *** LOADING DOMAIN 0 *** (XEN) **************************************** (XEN) Panic on CPU 0: (XEN) crc error (XEN) **************************************** Can you give me a light? Thanks Lei -- "We learn from failure, not from
2010 May 19
4
xen4.0.0 64 bit boot "panic on cpu 0, crc errror"
Hi experts, I boot the xen4.0.0 with gpxe and boot xen 32 bit works well but when I try boot xen 64 bit, the error happens (XEN)Brought up 16 cpus (XEN) *** LOADING DOMAIN 0 *** (XEN) **************************************** (XEN) Panic on CPU 0: (XEN) crc error (XEN) **************************************** Can you give me a light? Thanks Lei -- "We learn from failure, not from
2018 Sep 14
1
Re: NUMA issues on virtualized hosts
Hello again, when the iozone writes slow. This is how slabtop looks like: 62476752 62476728 0% 0.10K 1601968 39 6407872K buffer_head 1000678 999168 0% 0.56K 142954 7 571816K radix_tree_node 132184 125911 0% 0.03K 1066 124 4264K kmalloc-32 118496 118224 0% 0.12K 3703 32 14812K kmalloc-node 73206 56467 0% 0.19K 3486 21
2010 May 18
7
How to build a 64 bit xen
Hi experts,'' I build xen on my 64 bit redhat with "make xen", I got the image xen-4.0.0.gz then $gunzip xen-4.0.0.gz $file xen-4.0.0 it tells me it''s 32 bit,but I saw it link the 64 bit lib Thanks Lei -- "We learn from failure, not from success!" _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com
2010 May 18
7
How to build a 64 bit xen
Hi experts,'' I build xen on my 64 bit redhat with "make xen", I got the image xen-4.0.0.gz then $gunzip xen-4.0.0.gz $file xen-4.0.0 it tells me it''s 32 bit,but I saw it link the 64 bit lib Thanks Lei -- "We learn from failure, not from success!" _______________________________________________ Xen-devel mailing list Xen-devel@lists.xensource.com
2010 Jul 14
3
Do we have a tree to support 2.6.34 dom0 kernel
Hi Pasik, Do we have a way to build a 2.6.34 kernel as dom0 kernel can you show me the way? Thanks Lei _______________________________________________ Xen-users mailing list Xen-users@lists.xensource.com http://lists.xensource.com/xen-users
2010 Feb 17
3
processor affinity
Hello, i am running CentOS 5.4. i have a requirement where i need to have 1 application have a single processor all to its self, and the rest of the system run on the other processors. "taskman" lets me bind the process to a processor(s), but it does not make it exclusive. Is this possible to do? i have even tried mucking around with the rc.sysinit, but to no avail. thank you very
2013 Nov 27
1
glusterfs poor performance for my local machine (host and client on the same machine)
Hi experts I have a machine which have 5 hard disk I want to use glusterfs to fast my disk my machine's IP is 123.224.178.67 my steps 1)create the volume gluster volume create vol1 123.224.178.67:/buildarea2 123.224.178.67:/buildarea3 123.224.178.67:/buildarea4 123.224.178.67:/buildarea5 123.224.178.67: /buildarea6 2)start the volume gluster volume start vol1 3)mount the volume
2018 Sep 18
1
Re: NUMA issues on virtualized hosts
On 09/17/2018 04:59 PM, Lukas Hejtmanek wrote: > Hello, > > so the current domain configuration: > <cpu mode='host-passthrough'><topology sockets='8' cores='4' threads='1'/><numa><cell cpus='0-3' memory='62000000' /><cell cpus='4-7' memory='62000000' /><cell cpus='8-11'