On Fri, Aug 3, 2012 at 1:37 PM, Salvatore Maltese <inf.maltese@gmail.com>
wrote:> Sorry iam newb on virtualization , I must use Xen for a project of
> Distributed and Cooperative backup application and I would simulate it with
> virtual machines hence with XEN. Unfortunally iam tring since two week for
> install it . I would know something form started , because iam becoming
flip
> . I would use as Domain 0 linux mint 13 32 bit .
> i had seen some package , but i am not be able to start by boot . please
> help me . And sorry for my terrible English. thx
Do you mean your host (dom0) cannot boot? Are you using packages from
Linux Mint 13?
The best thing is to choose a distribution that has decent Xen support
out-of-the-box. Debian squeeze and Ubuntu 12.04 are both good
options; I think recent Fedoras have good Xen support as well.
-George
>
>
> On Fri, Aug 3, 2012 at 2:00 PM, <xen-users-request@lists.xen.org>
wrote:
>>
>> Send Xen-users mailing list submissions to
>> xen-users@lists.xen.org
>>
>> To subscribe or unsubscribe via the World Wide Web, visit
>> http://lists.xen.org/cgi-bin/mailman/listinfo/xen-users
>> or, via email, send a message with subject or body
''help'' to
>> xen-users-request@lists.xen.org
>>
>> You can reach the person managing the list at
>> xen-users-owner@lists.xen.org
>>
>> When replying, please edit your Subject line so it is more specific
>> than "Re: Contents of Xen-users digest..."
>>
>>
>> Today''s Topics:
>>
>> 1. Re: any opensource controllers to recommand? (Joseph Glanville)
>> 2. Re: XEN HA Cluster with LVM fencing and live migration ? The
>> right way ? (Joseph Glanville)
>> 3. settinjg dom0 memory, where did I go wrong? (Michael Egglestone)
>> 4. Re: settinjg dom0 memory, where did I go wrong?
>> (Emmanuel COURCELLE)
>> 5. Shoehorning a domU and "missing" memory (Xen 4.0.1)
(Andrew Wade)
>> 6. Re: Transcendent Memory ("tmem") -capable kernel now
publicly
>> released (gavin gao)
>> 7. Re: Transcendent Memory ("tmem") -capable kernel now
publicly
>> released (Stephan Seitz)
>>
>>
>> ----------------------------------------------------------------------
>>
>> Message: 1
>> Date: Fri, 3 Aug 2012 10:07:49 +1000
>> From: Joseph Glanville <joseph.glanville@orionvm.com.au>
>> To: David Erickson <halcyon1981@gmail.com>
>> Cc: yue wang <heuye.wang@gmail.com>, xen-users@lists.xen.org
>> Subject: Re: [Xen-users] any opensource controllers to recommand?
>> Message-ID:
>>
>>
<CAOzFzEjOtm2YFZO6frKjH8bLwSUYfiCL0F6Yc2=2iF-9VTWebw@mail.gmail.com>
>> Content-Type: text/plain; charset="utf-8"
>>
>> On 2 August 2012 10:39, David Erickson <halcyon1981@gmail.com>
wrote:
>>
>> > I am biased as the author of Beacon
(http://www.beaconcontroller.net/),
>> > but I have been using it with a cluster of 80 XenServer machines
running
>> > OVS, interconnected by physical OpenFlow switches for over a year.
>> >
>>
>> I''ll second Beacon as a good choice. Great performance and
probably the
>> most featureful of the open-source controllers.
>>
>> If you want to get hacking on stuff really fast then there is NOX which
>> though not as tanky as Beacon lets you write extensions in Python,
which
>> is
>> imo a big plus.
>>
>> There is also this project which I am yet to try out but it looks
>> interesting:
>>
>> https://github.com/trema
>>
>> It seems to be the remnants of the NEC Helios controller. (which to my
>> knowledge was never available anywhere).
>>
>>
>> >
>> > On Tue, Jul 31, 2012 at 7:57 PM, yue wang
<heuye.wang@gmail.com> wrote:
>> >
>> >> Hi, All
>> >>
>> >> do you have any opensource controllers to recommand?
>> >> since XCP don''t have a controller to control OVS
centrally,i need a
>> >> controller like vswitch controller for xenserver.
>> >> there are so many open source openflow controller, i really
don''t know
>> >> which one to choose.
>> >>
>> >> thanks in advance[?]
>> >>
>> >> _______________________________________________
>> >> Xen-users mailing list
>> >> Xen-users@lists.xen.org
>> >> http://lists.xen.org/xen-users
>> >>
>> >
>> >
>> > _______________________________________________
>> > Xen-users mailing list
>> > Xen-users@lists.xen.org
>> > http://lists.xen.org/xen-users
>> >
>>
>>
>>
>> --
>> CTO | Orion Virtualisation Solutions | www.orionvm.com.au
>> Phone: 1300 56 99 52 | Mobile: 0428 754 846
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL:
>>
<http://lists.xen.org/archives/html/xen-users/attachments/20120803/80feeefd/attachment.html>
>> -------------- next part --------------
>> A non-text attachment was scrubbed...
>> Name: 360.gif
>> Type: image/gif
>> Size: 453 bytes
>> Desc: not available
>> URL:
>>
<http://lists.xen.org/archives/html/xen-users/attachments/20120803/80feeefd/attachment.gif>
>>
>> ------------------------------
>>
>> Message: 2
>> Date: Fri, 3 Aug 2012 10:14:23 +1000
>> From: Joseph Glanville <joseph.glanville@orionvm.com.au>
>> To: Herve Roux <vevroux@msn.com>
>> Cc: xen-users@lists.xen.org
>> Subject: Re: [Xen-users] XEN HA Cluster with LVM fencing and live
>> migration ? The right way ?
>> Message-ID:
>>
>>
<CAOzFzEgvYFqyFKVCHDXKMvSiwx1Lfq0M6aQgAS4V3zJfZ=oPfg@mail.gmail.com>
>> Content-Type: text/plain; charset=windows-1252
>>
>> On 2 August 2012 18:19, Herve Roux <vevroux@msn.com> wrote:
>> > Hi,
>> >
>> >
>> >
>> > I am trying to build a rock solid XEN High availability cluster.
The
>> > platform is SLES 11 SP1 running on 2 HP DL585 both connected
through HBA
>> > fiber channel to the SAN (HP EVA).
>> >
>> > XEN is running smoothly and I?m even amazed with the live
migration
>> > performances (this is the first time I have the chance to try it
in such
>> > a
>> > nice environment).
>> >
>> > XEN apart the SLES heartbeat cluster is running fine as well and
they
>> > both
>> > interact nicely.
>> >
>> > Where I?m having some doubts is regarding the storage layout. I
have
>> > tried
>> > several configurations but each time I have to compromise. And
here is
>> > the
>> > problem, I don?t like to compromise ;)
>> >
>> >
>> >
>> > First I?ve tried to use a SAN LUN per Guest (using directly the
>> > multipath dm
>> > device as phy disk ). This is working nicely, live migration works
fine,
>> > easy setup even if the multipath.conf can get a bit fussy with the
>> > growing
>> > number of LUNs : but no fencing at all, I can start the VM on both
node
>> > and
>> > this is BAD!
>> >
>> > Then I?ve tried to used cLVM on top of the multipath. I?ve managed
to
>> > get
>> > cLVM up and running pretty easily in the cluster environment.
>> >
>> > From here to way of thinking:
>> >
>> > 1. One big SR on the SAN split into LV that I can use for my
VM. A
>> > huge step forward flexibility, no need to reconfigure the SAN each
time?
>> > Still with this solution the SR VG is open in shared mode between
the
>> > nodes
>> > and I don?t have low level lock of the storage. I can start a VM
two
>> > time
>> > and this is bad bad bad?
>> >
>> > 2. In order to provide fencing at the LVM level I can take
another
>> > approach: 1 VG per volume an open it in exclusive mode. The volume
will
>> > be
>> > active on one node at a time and I have no risk of data
corruption. The
>> > cluster will be in charge of balancing to volume when migrating VM
from
>> > one
>> > node to the other. But here the live migration is not working,
and this
>> > S?
>> >
>> >
>> >
>> > I was wondering what approach others have taken and if they is
something
>> > I?m
>> > missing.
>> >
>> > I?ve looked into the XEN locking system but from my point of view
the
>> > risk
>> > of dead lock is not ideal as well. From my point of view a DLM XEN
>> > locking
>> > system will be a good one, I don?t know if some work have been
done in
>> > the
>> > domain?
>> >
>> >
>> >
>> > Thanks in advance
>> >
>> > Herve
>> >
>> >
>> >
>> >
>> >
>> >
>> > _______________________________________________
>> > Xen-users mailing list
>> > Xen-users@lists.xen.org
>> > http://lists.xen.org/xen-users
>>
>> Personally I would steer clear of cLVM.
>>
>> If your SAN provides nice programatic control you can possibly
>> integrate it''s ability to do fencing into Pacemaker/Linux-HA
stacking
>> using a custom OCF (not a huge amount of work usually).
>> This would give you everything you want but is mainly determined by
>> how "open" your SAN is.
>>
>> Alternatively you can run with no storage layer fencing and make sure
>> you have proper STONITH in play.
>> This is pretty easy with Pacemaker/Corosync stack and can be done in
>> alot of ways. If you are pretty sure of your stacks stability (no
>> deadlocks/kernel oops etc) then you can just use SSH STONITH.
>> However if you want to be really damn sure then you can use IP PDU,
>> USB poweroff etc.
>>
>> It''s all about how sure you want to be/how much money/time you
have.
>>
>> Joseph.
>> --
>> CTO | Orion Virtualisation Solutions | www.orionvm.com.au
>> Phone: 1300 56 99 52 | Mobile: 0428 754 846
>>
>>
>>
>> ------------------------------
>>
>> Message: 3
>> Date: Thu, 02 Aug 2012 21:48:22 -0700
>> From: "Michael Egglestone" <mike@estone.ca>
>> To: xen-users@lists.xen.org
>> Subject: [Xen-users] settinjg dom0 memory, where did I go wrong?
>> Message-ID:
<fc.00000001e6690c0f00000001e6690c0f.e6690c10@estone.ca>
>> Content-Type: text/plain; charset="utf-8"
>>
>> Hello,
>> I''m trying to set dom0 to 4G of memory.
>> (Let me quickly say I don''t know if 4G of RAM for dom0 is good
idea, but I
>> thought I would try it, please advise otherwise) :)
>> My system has 32Gig of memory. (Xeon''s with Debian 6.0.5)
>>
>> Here is my /etc/default/grub
>>
>> GRUB_CMDLINE_XEN_DEFAULT="dom0_mem=4096M dom0_vcpus_pin"
>>
>> Here is my /etc/xen/xend-config.sxp
>>
>> [snip]
>> # dom0-min-mem is the lowest permissible memory level (in MB) for dom0.
>> # This is a minimum both for auto-ballooning (as enabled by
>> # enable-dom0-ballooning below) and for xm mem-set when applied to
dom0.
>> (dom0-min-mem 196)
>>
>> # Whether to enable auto-ballooning of dom0 to allow domUs to be
created.
>> # If enable-dom0-ballooning = no, dom0 will never balloon out.
>> (enable-dom0-ballooning no)
>>
>> # 32-bit paravirtual domains can only consume physical
>> # memory below 168GB. On systems with memory beyond that address,
>> # they''ll be confined to memory below 128GB.
>> # Using total_available_memory (in GB) to specify the amount of memory
>> reserved
>> # in the memory pool exclusively for 32-bit paravirtual domains.
>> # Additionally you should use dom0_mem = <-Value> as a parameter
in
>> # xen kernel to reserve the memory for 32-bit paravirtual domains,
default
>> # is "0" (0GB).
>> (total_available_memory 0)
>>
>> # In SMP system, dom0 will use dom0-cpus # of CPUS
>> # If dom0-cpus = 0, dom0 will take all cpus available
>> (dom0-cpus 4)
>> [snip]
>>
>> I''ve updated grub to populate /boot/grub/grub.cfg and then
rebooted.
>> It boots, and then I run top on my dom0 which shows this:
>>
>> Tasks: 153 total, 1 running, 152 sleeping, 0 stopped, 0 zombie
>> Cpu(s): 0.1%us, 0.3%sy, 0.0%ni, 96.9%id, 1.8%wa, 0.0%hi, 0.0%si,
>> 0.9%st
>> Mem: 2267508k total, 1641848k used, 625660k free, 88140k
buffers
>> Swap: 0k total, 0k used, 0k free, 1279128k cached
>>
>> Why do I only have 2Gig of RAM?
>>
>> Here are my domU''s...
>>
>> root@xen:/etc/xen# xm list
>> Name ID Mem VCPUs State
>> Time(s)
>> Domain-0 0 2557 4 r-----
>> 252.0
>> debian-central 1 1024 2 -b----
>> 22.2
>> debian-cms 2 4096 2 -b----
>> 186.7
>> debian-ldap 3 1024 2 -b----
>> 4.6
>> debian-ts 4 1024 2 -b----
>> 42.8
>> redhat-sdsweb 5 4096 4 -b----
>> 58.7
>> w2k3-awards 6 2048 2 -b----
>> 225.4
>> w2k3-sme 8 2048 2 -b----
>> 33.5
>> w2k8-sme 7 2048 2 -b----
>> 52.6
>> root@xen:/etc/xen#
>>
>> root@xen:/etc/xen# xm info
>> host : xen.sd57.bc.ca
>> release : 2.6.32-5-xen-amd64
>> version : #1 SMP Sun May 6 08:57:29 UTC 2012
>> machine : x86_64
>> nr_cpus : 24
>> nr_nodes : 2
>> cores_per_socket : 6
>> threads_per_core : 2
>> cpu_mhz : 2800
>> hw_caps :
>> bfebfbff:2c100800:00000000:00001f40:029ee3ff:00000000:00000001:00000000
>> virt_caps : hvm hvm_directio
>> total_memory : 32704
>> free_memory : 12115
>> node_to_cpu : node0:0,2,4,6,8,10,12,14,16,18,20,22
>> node1:1,3,5,7,9,11,13,15,17,19,21,23
>> node_to_memory : node0:2172
>> node1:9942
>> node_to_dma32_mem : node0:2172
>> node1:0
>> max_node_id : 1
>> xen_major : 4
>> xen_minor : 0
>> xen_extra : .1
>> xen_caps : xen-3.0-x86_64 xen-3.0-x86_32p hvm-3.0-x86_32
>> hvm-3.0-x86_32p hvm-3.0-x86_64
>> xen_scheduler : credit
>> xen_pagesize : 4096
>> platform_params : virt_start=0xffff800000000000
>> xen_changeset : unavailable
>> xen_commandline : placeholder dom0_mem=4096M dom0_vcpus_pin
>> cc_compiler : gcc version 4.4.5 (Debian 4.4.5-8)
>> cc_compile_by : fw
>> cc_compile_domain : deneb.enyo.de
>> cc_compile_date : Thu Jun 21 06:41:09 UTC 2012
>> xend_config_format : 4
>> root@xen:/etc/xen#
>>
>> Thanks for your advice!
>>
>> Cheers,
>> Mike
>>
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL:
>>
<http://lists.xen.org/archives/html/xen-users/attachments/20120802/1c27f98b/attachment.html>
>>
>> ------------------------------
>>
>> Message: 4
>> Date: Fri, 03 Aug 2012 08:49:39 +0200
>> From: Emmanuel COURCELLE <emmanuel.courcelle@toulouse.inra.fr>
>> To: xen-users@lists.xen.org
>> Subject: Re: [Xen-users] settinjg dom0 memory, where did I go wrong?
>> Message-ID: <501B7483.6030908@toulouse.inra.fr>
>> Content-Type: text/plain; charset="iso-8859-1";
Format="flowed"
>>
>> Le 03/08/2012 06:48, Michael Egglestone a ?crit :
>> > Hello,
>> > I''m trying to set dom0 to 4G of memory.
>> > (Let me quickly say I don''t know if 4G of RAM for dom0 is
good idea,
>> > but I thought I would try it, please advise otherwise) :)
>> > My system has 32Gig of memory. (Xeon''s with Debian 6.0.5)
>> >
>> Hello
>>
>> I think you''re running Debian ? Which version ?
>>
>> We installed recently a server with 256Gb memory, Debian testing/xen
>> 4.1/kernel 3.3.4 (downloaded from kernel.org) to be able to get a guest
>> with as more as 200Gb memory, and all this stuff works with as low as
>> 512Mb for dom0 !
>>
>> However, top shows KiB Mem: 354520, free shows 354520 also,
>>
>> BUT
>>
>> xm top shows 523912K
>>
>> As far as I understand, xm top (or the equivalent if you''r
using the xl
>> stack) is a better tool than top and others for monitoring dom0.
>>
>> Sincerely,
>>
>> --
>> Emmanuel COURCELLE emmanuel.courcelle@toulouse.inra.fr
>> L.I.P.M. (UMR CNRS-INRA 2594/441) tel (33) 5-61-28-54-50
>> I.N.R.A. - 24 chemin de Borde Rouge - Auzeville
>> CS52627 - 31326 CASTANET TOLOSAN Cedex - FRANCE
>>
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL:
>>
<http://lists.xen.org/archives/html/xen-users/attachments/20120803/ab169856/attachment.html>
>>
>> ------------------------------
>>
>> Message: 5
>> Date: Fri, 03 Aug 2012 11:33:23 +0100
>> From: Andrew Wade <andrew@memset.com>
>> To: xen-users@lists.xen.org
>> Subject: [Xen-users] Shoehorning a domU and "missing" memory
(Xen
>> 4.0.1)
>> Message-ID: <501BA8F3.6060701@memset.com>
>> Content-Type: text/plain; charset=ISO-8859-1
>>
>> Hi,
>>
>> I''m seeing an issue with 4.0.1 on Debian 6 regarding
"missing" memory.
>>
>> My set up:
>>
>> * Xen 4.0.1 on Debian 6.0.5
>> * dom0-min-mem 512, enable-dom0-ballooning no, dom0_mem=512M (GRUB
>> config). (I also tried with dom0-min-mem 0)
>> * Host server has 32GB RAM
>>
>> # xm list
>> Name ID Mem VCPUs State
>> Time(s)
>> Domain-0 0 501 8 r-----
>> 12.0
>>
>> # xm info|grep memory
>> total_memory : 32758
>> free_memory : 31832
>>
>> 32758 (total) - 31832 (''free_memory'') - 512 (dom0) =
414MB unaccounted.
>> (No domus are running)
>>
>> I created an HVM domU with 31488 MB RAM (N.B. this is less than 31832
>> reported by xm info free_memory plus I calculated about ~250MB memory
>> overhead) and 4 VCPUs but it wouldn''t start due to
insufficient memory
>> available. I expected it to fit.
>>
>> Is there an official calculation for the memory overhead (for
>> tables/caches etc)?
>>
>> Can anyone explain why a domu with 31,488MB won''t start when
31,832MB is
>> free? I''m trying to calculate the most amount of RAM a domU
can have
>> (i.e. to occupy an entire server)
>>
>> Thanks.
>>
>> --
>> Andrew Wade
>>
>> Memset Ltd., registration number 4504980. 25 Frederick Sanger Road,
>> Guildford, Surrey, GU2 7YD, UK.
>>
>>
>>
>> ------------------------------
>>
>> Message: 6
>> Date: Fri, 3 Aug 2012 04:23:15 -0700 (PDT)
>> From: gavin gao <gavin20112012@yahoo.com.cn>
>> To: xen-users@lists.xensource.com
>> Subject: Re: [Xen-users] Transcendent Memory ("tmem")
-capable kernel
>> now publicly released
>> Message-ID: <1343992995864-5710502.post@n5.nabble.com>
>> Content-Type: text/plain; charset=us-ascii
>>
>>
>>
>> Hi everyone
>> This feature is very cool.When I execute "make linux kernel"
on my
>> VM(vcpus=4,memory=256M to simulate memory pressure),it take almost 30
>> hours,
>> and tmem reduce this time to 2 hours~~~
>>
>> I am go on testing it~~~
>>
>>
>>
>> Gavin
>>
>>
>>
>> --
>> View this message in context:
>>
http://xen.1045712.n5.nabble.com/Transcendent-Memory-tmem-capable-kernel-now-publicly-released-tp5587302p5710502.html
>> Sent from the Xen - User mailing list archive at Nabble.com.
>>
>>
>>
>> ------------------------------
>>
>> Message: 7
>> Date: Fri, 3 Aug 2012 11:47:22 +0000
>> From: Stephan Seitz <s.seitz@netzhaut.de>
>> To: gavin gao <gavin20112012@yahoo.com.cn>
>> Cc: "xen-users@lists.xensource.com"
<xen-users@lists.xensource.com>
>> Subject: Re: [Xen-users] Transcendent Memory ("tmem")
-capable kernel
>> now publicly released
>> Message-ID: <1343994442.9254.2.camel@wotan2>
>> Content-Type: text/plain; charset="utf-8"
>>
>> Hi,
>>
>> how are you implementing it?
>>
>> I''m currently using it on test machines with
>> xen bootargs tmem tmem_dedup tmem_compress
>> and
>> dom0 kernel bootargs tmem
>>
>> the domUs also have recent kernels with tmem bootargs as well as the
>> zcache module loaded.
>>
>> I noticed very high memory latency when overcommiting memory.
>>
>> cheers,
>>
>> Stephan
>>
>> -------------- next part --------------
>> An HTML attachment was scrubbed...
>> URL:
>>
<http://lists.xen.org/archives/html/xen-users/attachments/20120803/40f47ca8/attachment.html>
>> -------------- next part --------------
>> A non-text attachment was scrubbed...
>> Name: signature.asc
>> Type: application/pgp-signature
>> Size: 490 bytes
>> Desc: This is a digitally signed message part
>> URL:
>>
<http://lists.xen.org/archives/html/xen-users/attachments/20120803/40f47ca8/attachment.pgp>
>>
>> ------------------------------
>>
>> _______________________________________________
>> Xen-users mailing list
>> Xen-users@lists.xen.org
>> http://lists.xen.org/xen-users
>>
>>
>> End of Xen-users Digest, Vol 90, Issue 4
>> ****************************************
>
>
>
> _______________________________________________
> Xen-users mailing list
> Xen-users@lists.xen.org
> http://lists.xen.org/xen-users