On Wed, 25 Nov 2015, Gordon Messmer wrote:
>I really recommend using the fewest partitions possible,
>replacing
>a disk will require you to handle each partition individually.
This is not a large burden, but I do agree to keep it simple.
>It's probably best to do a small /boot on RAID1 and use the rest of the
>disk for a second RAID1 volume, with LVM on that.
The default configuration is a partition for /boot and one per drive for
LVM PVs all of which are used by a single VG, in which LVs for / (which
has a fixed upper size), swap (which uses a computed size) and if
there's space available /home (which uses all remaining space).
And there's the problem, the installer uses all available space leaving
none for a RAID of /boot/.
I believe this is due to an assumption that Red Hat makes, that their
customers are mostly enterprise and thus the drives that the installer
sees would be logical and backed by RAID.
I could wish that the installer would RAID /boot by default, which is
small enough that it is unlikely to be a burden for those with RAID
backed "drives" and a boon to those who do not.
What's needed is enough space for the installer to carve out additional
partitions on the additional drives. Which means manual partitioning,
and in particular you must delete or resize /home -- remember, this is
in the initial configuration so there's nothing to lose, and you can
create/resize it a little later anyway. If you have very small disks
you probably need to shrink / and/or swap.
Reduce the space to be used by enough for another /boot and you can
change /boot's device type to one of the RAIDs.
Reduce it to no more than what the first drive provides and you can also
setup RAID under the PV -- modify the volume group redundancy.
After which you can make mount points for anything else, e.g., /var/www,
and grow what you shrank.
If you like pain you can back into such a layout after installation is
complete, but I do not recommend it.
When there are more than 2 drives possible I prefer MD RAID for /boot
and LVM RAID for everything else -- the not likeable part is that even
though LVM RAID leverages MD RAID the status is not presented via
/proc/mdstat, which changes how one must monitor for failures. When
only 2 drives are possible odds are you want RAID1 and it might as well
be under the PV.
Using LVM RAID is interesting in that it allows you to decide later
without much effort the redundancy you want on a per mount point basis,
e.g., on a 4 drive system it is just two commands to create mount points
(LVs) using RAID10 for /var/lib/*sql but RAID5 for /var/www -- that's
not impossible to do with MD RAID under PVs but it probably means lots
more commands and even some juggling.
/mark