similar to: LXC vCPU controlling problem

Displaying 20 results from an estimated 9000 matches similar to: "LXC vCPU controlling problem"

2015 Aug 12
1
Libvirt LXC vcpu doesn't seem to work
Hi, I seem to have a problem when creating a LXC container through virsh. While virsh -c lxc:/// dominfo <container> shows up (for example) 2 VCPUs as defined, if I run a CPU intensive task (such as stress --cpu 10) it will max out 10 CPU cores on the host. If I echo "0" > /cgroup/cpuset/libvirt/lxc/<domain>/cpuset.cpus then the container is properly confined to just
2019 Jan 21
2
libvirt 5.0.0 - LXC container still in "virsh list" output after shutdown
Hello. Centos 7.6 with libvirt build from base "virt" repository: libvirt-daemon-driver-lxc-5.0.0-1.el7.x86_64 libvirt-client-5.0.0-1.el7.x86_64 libvirt-daemon-5.0.0-1.el7.x86_64 libvirt-daemon-driver-network-5.0.0-1.el7.x86_64 libvirt-libs-5.0.0-1.el7.x86_64 + systemd-219-62.el7_6.2.x86_64 Now lxc containers with type='direct' can be started, but can't be stopped :)
2014 Feb 15
2
lxc and cgroups
Dear list, I am bulding a VM using libvirt and lxc for linux container. I have an issue with my cgroups settings: gabx@hortensia ➤➤ ~ # virsh start dahlia error: Failed to start domain dahlia error: internal error: No valid cgroup for machine dahlia My environement: Host : Arch Linux - systemd 208-11, libvirt 1.2.1-4 guest : Arch Linux , machine name : dahlia , /etc/libvirt/lxc/dahlia.xml ,
2019 Jan 22
0
Re: libvirt 5.0.0 - LXC container still in "virsh list" output after shutdown
> Centos 7.6 with libvirt build from base "virt" repository: > libvirt-daemon-driver-lxc-5.0.0-1.el7.x86_64 > libvirt-client-5.0.0-1.el7.x86_64 > libvirt-daemon-5.0.0-1.el7.x86_64 > libvirt-daemon-driver-network-5.0.0-1.el7.x86_64 > libvirt-libs-5.0.0-1.el7.x86_64 > + > systemd-219-62.el7_6.2.x86_64 > > Now lxc containers with type='direct' can be
2013 Feb 08
1
libvirt v1.0.2 fails to boot LXC container, but v1.0.0 works
Hello. tl;dr = v1.0.0 can boot my LXC containers, v.1.0.1 and v.1.0.2 fails. Paraphrased error message: "lxcContainerMountProcFuse:616 : Failed to mount ..../meminfo" I'd like to know if my host is misconfigured, or my domains, or ... why 1.0.2 and 1.0.1 are not working for me. I've been using libvirt for a while to manage QEMU instances. I have experimented with lxc.
2013 Aug 07
2
Is there any virsh command to setup cpusettune for lxc?
Hi Gao feng, I noticed one of your patch which adds cpuset cgroup support for lxc have been merged in libvirt 1.0.4. But I can't find any virsh command to set cpusettune for lxc container. Is there anyone? And how can I configure cpusettune for lxc container lively? Thanks ------------------ Best regards! GuanQiang
2012 Nov 08
1
OpenStack+libvirt+lxc: lxcContainerGetSubtree:1199 : Failed to read /proc/mounts
Hi, I'm running OpenStack on CentOS 6.3 to manage lxc instances. And running into series of problem relating libvirt and lxc interaction. For example, libvirt_lxc segfault ( https://bugzilla.redhat.com/show_bug.cgi?id=874549) which has an upstream fix. And another bugs such as fail to start when SELinux disabled. Finally, I decides to adopt libvirt 0.10.2, self compiled from
2012 Nov 07
1
libvirt with lxc: internal error The 'cpuacct', 'devices' & 'memory' cgroups controllers must be mounted
Hi, I'm running OpenStack to manage LXC instance through libvirt. The same setting runs perfectly well on Ubuntu 12.04, while on CentOS 6u3, libvirt dumped following message when starting lxc instance. virsh -c lxc:/// start instance-00000032 error: internal error The 'cpuacct', 'devices' & 'memory' cgroups controllers must be mounted Really appreciated if
2017 Feb 03
0
How to pin an libvirt lxc to a specific physical CPU core in openstack?
Dear all, In my applications, the real-time performance is very important, so I used 4 containers, with only one application running in each container and a physical CPU core is only dedicated to one lxc. and I must know which container is hosted by which CPU core. That is, I need to pin a specific lxc to a specific CPU core. I can achieve this in a native linux system by the following xml
2019 Sep 15
3
virsh -c lxc:/// setvcpus and <vcpu> configuration fails
Hi folks! i created a server with this XML file: <domain type='lxc'> <name>lxctest1</name> <uuid>227bd347-dd1d-4bfd-81e1-01052e91ffe2</uuid> <metadata> <libosinfo:libosinfo xmlns:libosinfo="http://libosinfo.org/xmlns/libvirt/domain/1.0"> <libosinfo:os id="http://centos.org/centos/6.9"/>
2014 Jan 15
0
Re: Does libvirt lxc driver support "cpuset" attribute?
On Wed, Jan 15, 2014 at 05:49:23AM +0000, WANG Cheng D wrote: > Dear all > > I allocate only one vcpu for the container by the following > statement, that is, I want to pin the vcpu to physical core "2". > <vcpu placement='static' cpuset="2" >1</vcpu> > My host has 4 physical cores. Before test, all the 4 cores are > idle. After I run 4
2013 Aug 07
0
Re: Is there any virsh command to setup cpusettune for lxc?
Hi Guan Qiang, On 08/07/2013 05:04 PM, hzguanqiang@corp.netease.com wrote: > Hi Gao feng, > > I noticed one of your patch which adds cpuset cgroup support for lxc have been merged in libvirt 1.0.4. > But I can't find any virsh command to set cpusettune for lxc container. > Is there anyone? And how can I configure cpusettune for lxc container lively? > numatune doesn't
2014 Dec 22
0
Re: why CPU pinning doesn't take effect when using lxc-enter-namespace to run an application
On Mon, Dec 22, 2014 at 08:40:03AM +0000, WANG Cheng D wrote: >Dear all, > >I want my container to run on the third CPU core and I define this by the following xml scrpits: ><vcpu placement="static" cpuset="3">1</vcpu> > >When I run my application in a container terminal, I can see the application runs on the third core as expected. >When I run
2014 Jan 15
2
Does libvirt lxc driver support "cpuset" attribute?
Dear all I allocate only one vcpu for the container by the following statement, that is, I want to pin the vcpu to physical core "2". <vcpu placement='static' cpuset="2" >1</vcpu> My host has 4 physical cores. Before test, all the 4 cores are idle. After I run 4 processes in the container, I found all the 4 cores in the host are 100% used. That is, the
2013 Jul 12
0
Re: libvrtd-1.1.0 crashes when attempting to start some (but not all) LXC containers
The debug log ends with this: 2013-07-12 16:43:31.740+0000: 21365: debug : virCgroupMakeGroup:708 : Make group /machine/dwj-hfax-dev.libvirt-lxc 2013-07-12 16:43:31.740+0000: 21365: debug : virCgroupMakeGroup:729 : Make controller /sys/fs/cgroup/cpu/machine/dwj-hfax-dev.libvirt-lxc/ 2013-07-12 16:43:31.740+0000: 21365: debug : virCgroupMakeGroup:729 : Make controller
2013 Jul 12
2
Re: libvrtd-1.1.0 crashes when attempting to start some (but not all) LXC containers
Update: I am able to edit the XML in "dwj-hfax-dev" such that libvirtd no longer crashes, and edit the XML for "dwj-lnx-dev" such that it will crash. The presents of "<seclabel type='none'/>" near the bottom causes libvirtd to crash. I do not recall ever manually adding that to my domain. In any event, libvirtd should probably not crash due to the
2014 Dec 22
2
why CPU pinning doesn't take effect when using lxc-enter-namespace to run an application
Dear all, I want my container to run on the third CPU core and I define this by the following xml scrpits: <vcpu placement="static" cpuset="3">1</vcpu> When I run my application in a container terminal, I can see the application runs on the third core as expected. When I run my application using lxc-enter-namespace, the CPU pinning doesn't take effect, i.e.,
2017 Oct 18
2
Can we disable write to /sys/fs/cgroup tree inside container ?
Hi all Each lxc container on node have mounted tmpfs for cgroups tree: [root-inside-lxc@tst1 ~]# mount | grep cgroups cgroup on /sys/fs/cgroup/cpu,cpuacct type cgroup (rw,nosuid,nodev,noexec,relatime,cpuacct,cpu) cgroup on /sys/fs/cgroup/cpuset type cgroup (rw,nosuid,nodev,noexec,relatime,cpuset) cgroup on /sys/fs/cgroup/memory type cgroup (rw,nosuid,nodev,noexec,relatime,memory) cgroup on
2019 Sep 16
0
Re: virsh -c lxc:/// setvcpus and <vcpu> configuration fails
On Sun, Sep 15, 2019 at 12:21:08PM +0200, info@layer7.net wrote: >Hi folks! > >i created a server with this XML file: > ><domain type='lxc'> > <name>lxctest1</name> > <uuid>227bd347-dd1d-4bfd-81e1-01052e91ffe2</uuid> > <metadata> > <libosinfo:libosinfo
2013 Sep 03
2
No valid cgroup for machine...
Hello! How do i get pass this error? offlinehacker:~/ $ virsh --debug 0 -c lxc:/// create o1.xml create: file(optdata): o1.xml error: Failed to create domain from o1.xml error: internal error: No valid cgroup for machine c1 My cgroups seem to be mounted: cgroup on /sys/fs/cgroup/systemd type cgroup