I am planning to increase the disk space on my desktop system. It is running CentOS 5.9 w/XEN. I have two 160Gig 2.5" laptop (2.5") SATA drives in two slots of a 4-slot hot swap bay configured like this: Disk /dev/sda: 160.0 GB, 160041885696 bytes 255 heads, 63 sectors/track, 19457 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 * 1 125 1004031 fd Linux raid autodetect /dev/sda2 126 19457 155284290 fd Linux raid autodetect Disk /dev/sdb: 160.0 GB, 160041885696 bytes 255 heads, 63 sectors/track, 19457 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdb1 * 1 125 1004031 fd Linux raid autodetect /dev/sdb2 126 19457 155284290 fd Linux raid autodetect sauron.deepsoft.com% cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sdb1[1] sda1[0] 1003904 blocks [2/2] [UU] md1 : active raid1 sdb2[1] sda2[0] 155284224 blocks [2/2] [UU] unused devices: <none> That is I have two RAID1 arrays: a small (1Gig) one mounted as /boot and a larger 148Gig one that is a LVM Volume Group (which contains a pile of file systems, some for DOM0 and some that are for other VMs). What I plan on doing is getting a pair of 320Gig 2.5" (laptop) SATA disks and fail over the existing disks to this new pair. I believe I can then 'grow' the second RAID array to be like ~300Gig. My question is: what happens to the LVM Volume Group? Will it grow when the RAID array grows? Or should I leave /dev/md1 its current size and create a new RAID array and add this as a second PV and grow the Volume Group that way? The documentation is not clear as to what happens -- the VG is marked 'resisable'. sauron.deepsoft.com% sudo pvdisplay --- Physical volume --- PV Name /dev/md1 VG Name sauron PV Size 148.09 GB / not usable 768.00 KB Allocatable yes PE Size (KByte) 4096 Total PE 37911 Free PE 204 Allocated PE 37707 PV UUID ttB15B-3eWx-4ioj-TUvm-lAPM-z9rD-Prumee sauron.deepsoft.com% sudo vgdisplay --- Volume group --- VG Name sauron System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 65 VG Access read/write VG Status resizable MAX LV 0 Cur LV 17 Open LV 12 Max PV 0 Cur PV 1 Act PV 1 VG Size 148.09 GB PE Size 4.00 MB Total PE 37911 Alloc PE / Size 37707 / 147.29 GB Free PE / Size 204 / 816.00 MB VG UUID qG8gCf-3vou-7dp2-Ar0B-p8jz-eXZF-3vOONr -- Robert Heller -- 978-544-6933 / heller at deepsoft.com Deepwoods Software -- http://www.deepsoft.com/ () ascii ribbon campaign -- against html e-mail /\ www.asciiribbon.org -- against proprietary attachments
On Mon, Feb 4, 2013 at 6:40 PM, Robert Heller <heller at deepsoft.com> wrote:> I am planning to increase the disk space on my desktop system. It is > running CentOS 5.9 w/XEN. I have two 160Gig 2.5" laptop (2.5") SATA drives > in two slots of a 4-slot hot swap bay configured like this: > >I would certainly suggest testing these steps out in some form or another as a dry run. And verify your backups and/or create new ones ahead of time. :) Create your new softraid mirror with the larger disks, add that device (we'll say md3 - since the new /boot will be md2 temporarily) to your existing LVM volume group, then pvmove the physical extents from the existing disk /dev/md1 to /dev/md3. Then verify that the extents have been moved off md1 and are on md3. Finally remove md1 from the VG with pvremove. I don't recall if I specifically used these notes, but the process matches the one I used. [0] A few more bits of info I recognize from my readings a while back. [1] [2] [3] I expect you could easily tweak your mdadm config and rebuild your initial ramdisk so that the next time you reboot there isn't an issue. This all depends on how much of this process you plan on doing online. [0] http://www.rhcedan.com/2010/10/20/migrating-physical-volumes-in-a-lvm2-volume-group/ [1] http://www.centos.org/docs/5/html/Cluster_Logical_Volume_Manager/move_new_ex4.html [2] http://tldp.org/HOWTO/LVM-HOWTO/removeadisk.html [3] http://www.whmcr.com/2011/06/21/moving-part-of-an-lvm-vg-from-one-pv-to-another/ sauron.deepsoft.com% cat /proc/mdstat> Personalities : [raid1] > md0 : active raid1 sdb1[1] sda1[0] > 1003904 blocks [2/2] [UU] > > md1 : active raid1 sdb2[1] sda2[0] > 155284224 blocks [2/2] [UU] > > unused devices: <none> > > That is I have two RAID1 arrays: a small (1Gig) one mounted as /boot > and a larger 148Gig one that is a LVM Volume Group (which contains a > pile of file systems, some for DOM0 and some that are for other VMs). > What I plan on doing is getting a pair of 320Gig 2.5" (laptop) SATA > disks and fail over the existing disks to this new pair. I believe I > can then 'grow' the second RAID array to be like ~300Gig. My question > is: what happens to the LVM Volume Group? Will it grow when the RAID >In my experience the VG will grow by whatever extents the new PV has. I found it quite helpful to use loopback devices to test possible softraid and/or LVM scenarios.> array grows? Or should I leave /dev/md1 its current size and create a > new RAID array and add this as a second PV and grow the Volume Group >Just add md3 to the VG and move the extents as noted above.> that way? The documentation is not clear as to what happens -- the VG > is marked 'resisable'. >The biggest gotcha is with PVs. Example even though this doesn't apply: In that if you plan on enlarging a disk (ex: LVM backing for a VM that uses LVM), the PV in the VM has to be offline to resize (the OS doesn't recognize the larger disk until after a reboot, but cannot be resized since the PV is online again!). You avoid that pitfall since you will have a separate disk (md3) that can be hot added.> > sauron.deepsoft.com% sudo pvdisplay > --- Physical volume --- > PV Name /dev/md1 > VG Name sauron > PV Size 148.09 GB / not usable 768.00 KB > Allocatable yes > PE Size (KByte) 4096 > Total PE 37911 > Free PE 204 > Allocated PE 37707 > PV UUID ttB15B-3eWx-4ioj-TUvm-lAPM-z9rD-Prumee > > sauron.deepsoft.com% sudo vgdisplay > --- Volume group --- > VG Name sauron > System ID > Format lvm2 > Metadata Areas 1 > Metadata Sequence No 65 > VG Access read/write > VG Status resizable > MAX LV 0 > Cur LV 17 > Open LV 12 > Max PV 0 > Cur PV 1 > Act PV 1 > VG Size 148.09 GB > PE Size 4.00 MB > Total PE 37911 > Alloc PE / Size 37707 / 147.29 GB > Free PE / Size 204 / 816.00 MB > VG UUID qG8gCf-3vou-7dp2-Ar0B-p8jz-eXZF-3vOONr > > > > -- > Robert Heller -- 978-544-6933 / heller at deepsoft.com > Deepwoods Software -- http://www.deepsoft.com/ > () ascii ribbon campaign -- against html e-mail > /\ www.asciiribbon.org -- against proprietary attachments > > > > _______________________________________________ > CentOS mailing list > CentOS at centos.org > http://lists.centos.org/mailman/listinfo/centos >Hope this helps. Best Regards, -- ---~~.~~--- Mike // SilverTip257 //
On 02/04/2013 06:40 PM, Robert Heller wrote:> I am planning to increase the disk space on my desktop system. It is > running CentOS 5.9 w/XEN. I have two 160Gig 2.5" laptop (2.5") SATA drives > in two slots of a 4-slot hot swap bay configured like this: > > Disk /dev/sda: 160.0 GB, 160041885696 bytes > 255 heads, 63 sectors/track, 19457 cylinders > Units = cylinders of 16065 * 512 = 8225280 bytes > > Device Boot Start End Blocks Id System > /dev/sda1 * 1 125 1004031 fd Linux raid autodetect > /dev/sda2 126 19457 155284290 fd Linux raid autodetect > > Disk /dev/sdb: 160.0 GB, 160041885696 bytes > 255 heads, 63 sectors/track, 19457 cylinders > Units = cylinders of 16065 * 512 = 8225280 bytes > > Device Boot Start End Blocks Id System > /dev/sdb1 * 1 125 1004031 fd Linux raid autodetect > /dev/sdb2 126 19457 155284290 fd Linux raid autodetect > > sauron.deepsoft.com% cat /proc/mdstat > Personalities : [raid1] > md0 : active raid1 sdb1[1] sda1[0] > 1003904 blocks [2/2] [UU] > > md1 : active raid1 sdb2[1] sda2[0] > 155284224 blocks [2/2] [UU] > > unused devices:<none> > > That is I have two RAID1 arrays: a small (1Gig) one mounted as /boot > and a larger 148Gig one that is a LVM Volume Group (which contains a > pile of file systems, some for DOM0 and some that are for other VMs). > What I plan on doing is getting a pair of 320Gig 2.5" (laptop) SATA > disks and fail over the existing disks to this new pair. I believe I > can then 'grow' the second RAID array to be like ~300Gig. My question > is: what happens to the LVM Volume Group? Will it grow when the RAID > array grows?Not on its own, but you can grow it. I believe the recommended way to do the LVM volume is to partition new drive as type fd install new PV on new partition (will be new, larger size) make new PV part of old volume group migrate all volumes on old PV onto new PV remove old PV from volume group You have to do this separately for each drive, but it isn't very hard. Of course your boot partition will have to be handled separately.> Or should I leave /dev/md1 its current size and create a > new RAID array and add this as a second PV and grow the Volume Group > that way?That is a solution to a different problem. You would end up with a VG of about 450 GB total. If that is what you want to do, that works too.> The documentation is not clear as to what happens -- the VG > is marked 'resisable'. > > sauron.deepsoft.com% sudo pvdisplay > --- Physical volume --- > PV Name /dev/md1 > VG Name sauron > PV Size 148.09 GB / not usable 768.00 KB > Allocatable yes > PE Size (KByte) 4096 > Total PE 37911 > Free PE 204 > Allocated PE 37707 > PV UUID ttB15B-3eWx-4ioj-TUvm-lAPM-z9rD-Prumee > > sauron.deepsoft.com% sudo vgdisplay > --- Volume group --- > VG Name sauron > System ID > Format lvm2 > Metadata Areas 1 > Metadata Sequence No 65 > VG Access read/write > VG Status resizable > MAX LV 0 > Cur LV 17 > Open LV 12 > Max PV 0 > Cur PV 1 > Act PV 1 > VG Size 148.09 GB > PE Size 4.00 MB > Total PE 37911 > Alloc PE / Size 37707 / 147.29 GB > Free PE / Size 204 / 816.00 MB > VG UUID qG8gCf-3vou-7dp2-Ar0B-p8jz-eXZF-3vOONr >Doesn't look like anyone answered your question, so I'll tell you that the answer is "Yes". Ted Miller
On Thu, Feb 14, 2013 at 11:58 PM, Ted Miller <tedlists at sbcglobal.net> wrote:> On 02/04/2013 06:40 PM, Robert Heller wrote: > > I am planning to increase the disk space on my desktop system. It is > > running CentOS 5.9 w/XEN. I have two 160Gig 2.5" laptop (2.5") SATA > drives > > in two slots of a 4-slot hot swap bay configured like this: > > > > Disk /dev/sda: 160.0 GB, 160041885696 bytes > > 255 heads, 63 sectors/track, 19457 cylinders > > Units = cylinders of 16065 * 512 = 8225280 bytes > > > > Device Boot Start End Blocks Id System > > /dev/sda1 * 1 125 1004031 fd Linux raid > autodetect > > /dev/sda2 126 19457 155284290 fd Linux raid > autodetect > > > > Disk /dev/sdb: 160.0 GB, 160041885696 bytes > > 255 heads, 63 sectors/track, 19457 cylinders > > Units = cylinders of 16065 * 512 = 8225280 bytes > > > > Device Boot Start End Blocks Id System > > /dev/sdb1 * 1 125 1004031 fd Linux raid > autodetect > > /dev/sdb2 126 19457 155284290 fd Linux raid > autodetect > > > > sauron.deepsoft.com% cat /proc/mdstat > > Personalities : [raid1] > > md0 : active raid1 sdb1[1] sda1[0] > > 1003904 blocks [2/2] [UU] > > > > md1 : active raid1 sdb2[1] sda2[0] > > 155284224 blocks [2/2] [UU] > > > > unused devices:<none> > > > > That is I have two RAID1 arrays: a small (1Gig) one mounted as /boot > > and a larger 148Gig one that is a LVM Volume Group (which contains a > > pile of file systems, some for DOM0 and some that are for other VMs). > > What I plan on doing is getting a pair of 320Gig 2.5" (laptop) SATA > > disks and fail over the existing disks to this new pair. I believe I > > can then 'grow' the second RAID array to be like ~300Gig. My question > > is: what happens to the LVM Volume Group? Will it grow when the RAID > > array grows? > > Not on its own, but you can grow it. I believe the recommended way to do > the LVM volume is to > partition new drive as type fd >LVM is 8e Software RAID is fd> install new PV on new partition (will be new, larger size) > make new PV part of old volume group > migrate all volumes on old PV onto new PV > remove old PV from volume group > > You have to do this separately for each drive, but it isn't very hard. Of > course your boot partition will have to be handled separately. > >This is what I said ;) http://lists.centos.org/pipermail/centos/2013-February/131917.html> > > Or should I leave /dev/md1 its current size and create a > > new RAID array and add this as a second PV and grow the Volume Group > > that way? > > That is a solution to a different problem. You would end up with a VG of > about 450 GB total. If that is what you want to do, that works too. > >He has to leave /dev/md1 at its current size ... it's a raid1.> > The documentation is not clear as to what happens -- the VG > > is marked 'resisable'. > > > > sauron.deepsoft.com% sudo pvdisplay > > --- Physical volume --- > > PV Name /dev/md1 > > VG Name sauron > > PV Size 148.09 GB / not usable 768.00 KB > > Allocatable yes > > PE Size (KByte) 4096 > > Total PE 37911 > > Free PE 204 > > Allocated PE 37707 > > PV UUID ttB15B-3eWx-4ioj-TUvm-lAPM-z9rD-Prumee > > > > sauron.deepsoft.com% sudo vgdisplay > > --- Volume group --- > > VG Name sauron > > System ID > > Format lvm2 > > Metadata Areas 1 > > Metadata Sequence No 65 > > VG Access read/write > > VG Status resizable > > MAX LV 0 > > Cur LV 17 > > Open LV 12 > > Max PV 0 > > Cur PV 1 > > Act PV 1 > > VG Size 148.09 GB > > PE Size 4.00 MB > > Total PE 37911 > > Alloc PE / Size 37707 / 147.29 GB > > Free PE / Size 204 / 816.00 MB > > VG UUID qG8gCf-3vou-7dp2-Ar0B-p8jz-eXZF-3vOONr > > > Doesn't look like anyone answered your question, so I'll tell you that the > answer is "Yes". > > Ted Miller > > > _______________________________________________ > CentOS mailing list > CentOS at centos.org > http://lists.centos.org/mailman/listinfo/centos >-- ---~~.~~--- Mike // SilverTip257 //
Possibly Parallel Threads
- SATA DVD Burner / AHCI / CentOS 4.7 (kernel: 2.6.9-78.0.22.EL)
- USB Serial ports (ttyACMn) CentOS 6.7 (64-bit) vs. CentOS 6.7 (64-bit)
- Question about running a CentOS4.8 (32-bit) guest under CentOS5.4xen (64-bit)
- Best Distro for Mongrel
- page allocation failure