similar to: start lxc container on fedora 19

Displaying 20 results from an estimated 900 matches similar to: "start lxc container on fedora 19"

2015 Apr 08
4
Centos 7.1.1503 + libvirt 1.2.14 = broken direct network mode
Hi all. I use LXC on Centos 7 x86-64, with libvirt version 1.2.6 and 1.2.12 My container has bridged network: # virsh dumpxml test1 <domain type='lxc'> <name>test1</name> <uuid>518539ab-7491-45ab-bb1d-3d7f11bfb0b1</uuid> <memory unit='KiB'>1048576</memory> <currentMemory unit='KiB'>1048576</currentMemory>
2013 Jul 31
0
Re: start lxc container on fedora 19
On Wed, Jul 31, 2013 at 12:46:58PM +0530, Aarti Sawant wrote: > hello, > > i am new to lxc, i have created a lxc container on fedora 19 > i created a container rootfs of fedora 19 by using > yum --installroot=/containers/test1 --releasever=19 install openssh > > test1.xml file for container test1 > <domain type="lxc"> > <name>test1</name>
2014 Sep 15
2
cgroups inside LXC containers losts memory limits after some time
Hi all I have CentOS Linux release 7.0.1406, libvirt 1.2.7 installed. Just after create and start inside LXC container present cgroups. Example for memory: [root@ce7-t1 /]# ls -la /sys/fs/cgroup/memory/ total 0 drwxr-xr-x 2 root root 0 Sep 15 17:14 . drwxr-xr-x 12 root root 280 Sep 15 17:14 .. -rw-r--r-- 1 root root 0 Sep 15 17:14 cgroup.clone_children --w--w--w- 1 root root 0 Sep 15
2018 Aug 01
3
LXC Memory Limits wont work
Hello, iam currently trying to run LXC Containers with libvirt but the memory limit doesn't want to work in the container i see the full 32GB from the Host OS iam pretty sure that iam missing a configline in the xml lxc-template ~ # free -m total used free shared buff/cache available Mem: 32108 626 31396 249 85
2016 Nov 08
3
Sharing network namespace between containers
Hello Based on the lxc driver documentation, I am trying to create an xml to share an existing network namespace with another container. I am running libvirt 1.2.15. Here is the xml: <domain type='lxc' xmlns:lxc='http://libvirt.org/schemas/domain/lxc/1.0'> <name>nt</name> <uuid>43c00192-e114-4e29-8ce7-4b5487f60a75</uuid> <memory
2015 Apr 09
0
Re: Centos 7.1.1503 + libvirt 1.2.14 = broken direct network mode
On 04/08/2015 09:38 AM, mxs kolo wrote: > Hi all. > > I use LXC on Centos 7 x86-64, with libvirt version 1.2.6 and 1.2.12 > My container has bridged network: > # virsh dumpxml test1 > <domain type='lxc'> > <name>test1</name> > <uuid>518539ab-7491-45ab-bb1d-3d7f11bfb0b1</uuid> > <memory
2014 Sep 15
0
Re: cgroups inside LXC containers losts memory limits after some time
HI all >After unpredictable time passed (1-5 day ?), cgroups inside LXC >magicaly removed. virsh dumpxml config look like this: <domain type='lxc' id='3566'> <name>puppet</name> <uuid>6d49b280-5686-4e3c-b048-1b5d362fb137</uuid> <memory unit='KiB'>8388608</memory> <currentMemory
2014 Sep 16
2
1.2.7 and 1.2.8 fail to start container: libvirt_lxc[4904]: segfault at 0 ip ...error 4 in libc-2.17.so[
HI all Centos 7, 3.10.0-123.6.3.el7.x86_64 libvirt 1.27, libvirt 1.2.8 builded from source with ./configure --prefix=/usr make && make install LXC with direct network failed to start: Sep 16 19:19:38 node01 kernel: device br502 entered promiscuous mode Sep 16 19:19:39 node01 kernel: device br502 left promiscuous mode Sep 16 19:19:39 node01 avahi-daemon[1532]: Withdrawing workstation
2016 Nov 18
2
locking domain memory
Hi, is there a way to lock a guests memory so it doesn't get swapped out? I now there is memoryBacking->locked but that says it requires memtune->hard_limit and the description of that basically "don't ever do this" rendering the locked element kind of pointless. How can I prevent the guests memory from being swapped out without shooting myself in the foot? Regards,
2019 Apr 03
1
SEV machines and memory pinning
Hello, I am working on implementing SEV support in OpenStack. There are some questions that came up in the discussion of the spec [0] [0] https://review.openstack.org/#/c/641994/ As far as i understand, the memory for SEV machines need to be pinned so that it doesn't migrate to swap and page migration. ROMS, UEFI pflash and video RAM should be pinned too. Initially we planned to use
2019 Sep 15
3
virsh -c lxc:/// setvcpus and <vcpu> configuration fails
Hi folks! i created a server with this XML file: <domain type='lxc'> <name>lxctest1</name> <uuid>227bd347-dd1d-4bfd-81e1-01052e91ffe2</uuid> <metadata> <libosinfo:libosinfo xmlns:libosinfo="http://libosinfo.org/xmlns/libvirt/domain/1.0"> <libosinfo:os id="http://centos.org/centos/6.9"/>
2015 Feb 10
2
Re: HugePages - can't start guest that requires them
On 09.02.2015 18:19, G. Richard Bellamy wrote: > First I'll quickly summarize my understanding of how to configure numa... > > In "//memoryBacking/hugepages/page[@nodeset]" I am telling libvirt to > use hugepages for the guest, and to get those hugepages from a > particular host NUMA node. No, @nodeset refers to guest NUMA nodes. > > In
2016 Nov 22
1
Re: locking domain memory
On 21.11.2016 17:05, Michal Privoznik wrote: > On 18.11.2016 23:17, Dennis Jacobfeuerborn wrote: >> Hi, >> is there a way to lock a guests memory so it doesn't get swapped out? I >> now there is memoryBacking->locked but that says it requires >> memtune->hard_limit and the description of that basically "don't ever do >> this" rendering the
2011 Feb 19
1
Conditional recoding
I am trying to recode a variable into another variable and while the package 'car' works well when it is only recoding A into B, I am not sure I can do the same with recoding (A or C) into B. If i can use recode please advise on how to. So i am using an if/else if conditions. My sample dataset is below along with the code and the warning and results i get. TIA Krishnan #****Code****
2013 Nov 20
1
Failed to access the console after starting the lxc container
Hello, I am starting the system container without "/" directory So that it can share the host /bin /sbin /lib and /lib64 and i have sepeartely mounted /var, /etc and /usr directory for the container Below is xml file <domain type='lxc'> <name>test6</name> <memory>102400</memory> <os> <type>exe</type>
2015 May 22
2
libvirt with gcc5 Test failing
Hello! I'm trying to compile libvirt using GCC 5.1 but one of the test are failing and I have no idea why :( Hopefully someone of you could help me, here part of my log: ========================================== libvirt 1.2.14: tests/test-suite.log ========================================== # TOTAL: 109 # PASS: 107 # SKIP: 1 # XFAIL: 0 # FAIL: 1 # XPASS: 0 # ERROR: 0 .. contents::
2012 Oct 18
0
0.10.x incorrectly reporting currentMemory size
Hi, <memory unit='KiB'>16777216</memory> <currentMemory unit='KiB'>1394380</currentMemory> <memtune> <hard_limit unit='KiB'>1594380</hard_limit> <soft_limit unit='KiB'>1494380</soft_limit> </memtune> results to: 0.10.x, .dominfo. or dumpxml | grep -i currentmemory Max memory: 16777216
2013 Jul 12
2
libvrtd-1.1.0 crashes when attempting to start some (but not all) LXC containers
Hello all, I have two issues: 1) I am unable to start a seemingly correct LXC domain (I cloned it from a working domain). 2) I am able to crash "libvirtd" by attempting to start the cloned domain, but starting the original works just fine. I humbly submit that item #2 is a bug - the "libvirtd" daemon should never crash due to anything the "libvirt" client
2015 Feb 23
1
Re: HugePages - can't start guest that requires them
On 20.02.2015 21:32, G. Richard Bellamy wrote: > <snip/> > > I've modified my config [1] based on my understanding, and am running > into a new error. Basically I'm hitting the oom-killer [2] even though > the hard_limit [3] of memtune is below the total number of hugepages > set for that NUMA nodeset. > Just drop the hard_limit. It's a blackbox we should
2014 Mar 22
1
Issues to manage RAM on openvz guests
Hi all, I'm playing with virsh, and I succeed to mange easily some KVM nodes. But I have issues to size RAM of openvz guests. virsh set by default the the memory to 256M on each guest instead of that I specify in the XML I have to modify by end the variable PHYSPAGES on each /etc/vz/conf/id_of_my_vz.conf Please follow the XML dump and the steps that I did. Cheers,Aurelien Log ---- 1)