Displaying 20 results from an estimated 2000 matches similar to: "Domain vCPU threads affinity"
2018 Sep 05
0
Re: Domain vCPU threads affinity
On Wed, Sep 05, 2018 at 03:48:45PM +0300, Nikos Anastopoulos wrote:
>Hello,
>
>According to the docs, vcpupin will use either cgroups or sched_setaffinity
>to pin vcpu threads to cpus. How is this decision made?
>I observe differences even on different hosts featuring the same version of
>libvirtd (1.3.1): on one host vcpupin affects cpuset.cpus (cgroup), and on
>the other it
2018 Sep 17
2
Re: NUMA issues on virtualized hosts
On 09/14/2018 03:36 PM, Lukas Hejtmanek wrote:
> Hello,
>
> ok, I found that cpu pinning was wrong, so I corrected it to be 1:1. The issue
> with iozone remains the same.
>
> The spec is running, however, it runs slower than 1-NUMA case.
>
> The corrected XML looks like follows:
[Reformated XML for better reading]
<cpu mode="host-passthrough">
2018 Sep 14
1
Re: NUMA issues on virtualized hosts
Hello again,
when the iozone writes slow. This is how slabtop looks like:
62476752 62476728 0% 0.10K 1601968 39 6407872K buffer_head
1000678 999168 0% 0.56K 142954 7 571816K radix_tree_node
132184 125911 0% 0.03K 1066 124 4264K kmalloc-32
118496 118224 0% 0.12K 3703 32 14812K kmalloc-node
73206 56467 0% 0.19K 3486 21
2013 Dec 03
1
Re: help about migration with libvirt
On 12/03/2013 12:52 AM, Nicolas Sebrecht wrote:
> This is about CPU affinity. You may look at CPU pinning with 'vcpupin'
> in the manual of virsh. It is possible to enable vCPU pinning in libvirt
> guest XML.
>
> Don't know if libvirt supports applying CPU affinity at guest runtime
Yes, 'virsh vcpupin' can be used to change CPU affinity at runtime.
> but
2018 Sep 14
3
NUMA issues on virtualized hosts
Hello,
I have cluster with AMD EPYC 7351 cpu. Two CPUs per node. I have performance
8-NUMA configuration:
This is from hypervizor:
[root@hde10 ~]# lscpu
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
CPU(s): 64
On-line CPU(s) list: 0-63
Thread(s) per core: 2
Core(s) per socket: 16
Socket(s): 2
NUMA
2018 Sep 18
1
Re: NUMA issues on virtualized hosts
On 09/17/2018 04:59 PM, Lukas Hejtmanek wrote:
> Hello,
>
> so the current domain configuration:
> <cpu mode='host-passthrough'><topology sockets='8' cores='4' threads='1'/><numa><cell cpus='0-3' memory='62000000' /><cell cpus='4-7' memory='62000000' /><cell cpus='8-11'
2012 Apr 04
1
vcpupin and kvm
Hi.
Can anyone confirm if vcpupin works correctly with kvm? If set I pin a
vcpu it shows up correctly via vcpuinfo, but querying via taskset shows
the cpu is not pinned.
Thanks,
-Geoff
--
----
Geoff Galitz, ggalitz at shutterstock.com
WebOps Engineer, Europe
Shutterstock Images
http://wwww.shutterstock.com/
2013 Dec 03
3
help about migration with libvirt
Hi experts
I want to use libvirt doing the migration
migration the guest from core 0 to core 2, Is there a way to do this ? any
points?
Lei
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/libvirt-users/attachments/20131203/1cd982b2/attachment.htm>
2013 Jul 19
2
How to insert vcpupin in guest xml file
Hi all,
I am trying to add vcpupin in the guest xml file. I am working with openstack and the code I have is python bound. I investigated through the code and found that the elemets in xml file are set in get_guest_config function. Now the thing is I am not able to set vcpupin element. I mean I tried guest.cputune_vcpupin but it's not working.
Help me out, please!
Thanks.
~Peeyush Gupta
2017 Apr 26
3
Tunnelled migrate Windows7 VMs halted
[moderator note: I'm forwarding a stripped down version of the original
mail which was rejected in the moderator queue. I stripped the 3.3
megabyte .tar.bz2 of the log file attachment, which is inappropriate for
a technical list. Either trim the log to the relevant portion, or host
the log externally and have your list email merely give a URL of the
externally-hosted file]
>
2016 May 13
0
vcpupin vs taskset performance difference
Hi all,
I am trying to run some performance tests using libvirt 1.2.17. For these tests it is necessary to set a specific affinity for the VCPUs of a running VM.
The problem is, that I noticed a significant performance drop in my metrics when using 'vcpupin' from the virsh console instead of running 'taskset' natively.
Specifically the methods I am comparing are
1.
2014 Feb 12
2
Re: Help? Running into problems with migrateToURI2() and virDomainDefCheckABIStability()
On 02/11/2014 04:45 PM, Cole Robinson wrote:
> On 02/10/2014 06:46 PM, Chris Friesen wrote:
>> Hi,
>>
>> We've run into a problem with libvirt 1.1.2 and are looking for some comments
>> on whether this is a bug or design intent.
>>
>> We're trying to use migrateToURI() but we're using a few things (numatune,
>> vcpu mask, etc.) that may need
2013 Sep 06
21
[PATCH v2 0/5] xl: allow for node-wise specification of vcpu pinning
Hi all,
This is the second take of a patch that I submitted some time ago for allowing
specifying vcpu pinning taking NUMA nodes into account. IOW, something like
this:
* "nodes:0-3": all pCPUs of nodes 0,1,2,3;
* "nodes:0-3,^node:2": all pCPUS of nodes 0,1,3;
* "1,nodes:1-2,^6": pCPU 1 plus all pCPUs of nodes 1,2
but not pCPU 6;
v1 was a single patch, this is
2010 Feb 17
3
processor affinity
Hello,
i am running CentOS 5.4. i have a requirement where i need to have 1
application have a single processor all to its self, and the rest of the
system run on the other processors. "taskman" lets me bind the process
to a processor(s), but it does not make it exclusive. Is this possible
to do? i have even tried mucking around with the rc.sysinit, but to no
avail.
thank you very
2019 Sep 15
3
virsh -c lxc:/// setvcpus and <vcpu> configuration fails
Hi folks!
i created a server with this XML file:
<domain type='lxc'>
<name>lxctest1</name>
<uuid>227bd347-dd1d-4bfd-81e1-01052e91ffe2</uuid>
<metadata>
<libosinfo:libosinfo
xmlns:libosinfo="http://libosinfo.org/xmlns/libvirt/domain/1.0">
<libosinfo:os id="http://centos.org/centos/6.9"/>
2013 Jun 14
0
can virsh set the cpuset attribute of <vcpu ..> (CPU Allocation) ?
Is it possible to use virsh to set the cpuset attribute of the CPU
Allocation element in a domain ?
<domain>
...
<vcpu placement='static' cpuset="1-4,^3,6" current="1">2</vcpu>
...
</domain>
I have seen that virsh vcpupin and virsh emulatorpin can be used to query
and set
the cpusets of the <vcpupin> and <emulatorpin>
2014 Jan 15
2
Does libvirt lxc driver support "cpuset" attribute?
Dear all
I allocate only one vcpu for the container by the following statement, that is, I want to pin the vcpu to physical core "2".
<vcpu placement='static' cpuset="2" >1</vcpu>
My host has 4 physical cores. Before test, all the 4 cores are idle. After I run 4 processes in the container, I found all the 4 cores in the host are 100% used. That is, the
2013 Jul 19
2
pinVcpu not working
Hi all,
I am working with libvirt and I am trying to set cpu affinity. Now I can always use
virsh vcpupin <domain_name> <vcpu> <pcpu>
to pin vcpus to pcpus. I want to do it using Python API. Now, there is function pinVcpu which is supposed to do that. But this is not working. For example I gave
dom.pinVcpu(0,1)
but still my vcpu affinity is for all the pcpus. The
2010 Dec 06
1
R with ATLAS avoids Linux cpu affinity
Hi all,
I have a problem with cpu affinity in my R-2.11.1 installation compiled against ATLAS running on a Linux (Ubuntu 10.04) cluster under GridEngine. I wish to use Grid Engine's core binding feature to bind user processes into the number of cores they request on the cluster, thus preventing badly behaved multi-threaded libraries from consuming more cores than requested. An example of
2014 Feb 10
2
Help? Running into problems with migrateToURI2() and virDomainDefCheckABIStability()
Hi,
We've run into a problem with libvirt 1.1.2 and are looking for some
comments on whether this is a bug or design intent.
We're trying to use migrateToURI() but we're using a few things
(numatune, vcpu mask, etc.) that may need adjustment during the
migration. We found that migrateToURI2() mostly works if we use XML
created by copying the domain XML from the running instance