similar to: Software RAID woes due to GRUB

Displaying 20 results from an estimated 7000 matches similar to: "Software RAID woes due to GRUB"

2006 Oct 05
1
Cannot re-make a software raid pair
Apologies if you get this twice - the first one didn't seem to make it... Hi Guys, I have just replaced a faulty Max...woah, wait...this one's a Seagate... IDE hard disk but I cannot remake the software raid pair. The currently running disk is hda and I am trying to add back hdg - both are master drives on separate controllers. I have run fdisk on hdg and created the same partition
2013 Sep 15
1
grub command line
Hello Everyone I have a remote CentOS 6.4 server (with KVM access), when I received the server it was running with LVM on single disk (sda) I managed to remove LVM and install raid 1 in sda and sdb disks the mirroring is working fine, my only issue now is that everytime I reboot the server I got the grub command line and I have manually boot using comand grub> configfile
2008 Mar 23
4
md raid1 - no speed improvement
Hi, I have two 320 GB SATA disks (/dev/sda, /dev/sdb) in a server running CentOS release 5. They both have three partitions setup as RAID1 using md (boot, swap, and an LVM data partition). # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sdb1[1] sda1[0] 104320 blocks [2/2] [UU] md1 : active raid1 sdb2[1] sda2[0] 4192896 blocks [2/2] [UU] md2 : active raid1 sdb3[1]
2012 Jun 07
1
mdadm: failed to write superblock to
Hello, i have a little problem. Our server has an broken RAID. # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sda1[2](F) sdb1[1] 2096064 blocks [2/1] [_U] md2 : active raid1 sda3[2](F) sdb3[1] 1462516672 blocks [2/1] [_U] md1 : active raid1 sda2[0] sdb2[1] 524224 blocks [2/2] [UU] unused devices: <none> I have remove the partition: # mdadm --remove
2006 Apr 02
2
raid setup
Hi, I have 2 identical xSeries 346 with 2 identical IBM 72GB scsi drive. What i did is install the centos 4.2 serverCD on the first IBM and set the HDD to raid1 and raid0 for swap. Now what i did is get the 2nd HDD in the 1st Server swap it with the 1st HDD in the 2nd Server and rebuild the Raids. The 1st server rebuild the array fine. My problem is the Second server, after rebuilding it and
2011 Nov 19
2
extlinux from syslinux 4.04 failed hard after choosing bzImage.
Hi, Today I replaced grub(1) with syslinux on my laptop, after rebooting to syslinux and selecting kernel I got massive freeze along with colorful output to screen - http://i.imgur.com/lFjFc.jpg I think I tried everything, even doing esc and writing after boot: /bzImage-3.1.0, after enter I had the same freeze with the same screen artefacts. How I installed extlinux: mkdir /boot/extlinux
2014 Feb 07
3
Software RAID1 Failure Help
I am running software RAID1 on a somewhat critical server. Today I noticed one drive is giving errors. Good thing I had RAID. I planned on upgrading this server in next month or so. Just wandering if there was an easy way to fix this to avoid rushing the upgrade? Having a single drive is slowing down reads as well, I think. Thanks. Feb 7 15:28:28 server smartd[2980]: Device: /dev/sdb
2008 Oct 05
3
Software Raid Expert Needed
Hello all, I have 2 x 250GB sata disks (sda and sdb). # fdisk -l /dev/sda Disk /dev/sda: 250.0 GB, 250059350016 bytes 255 heads, 63 sectors/track, 30401 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 * 1 14939 119997486 fd Linux raid autodetect /dev/sda2 14940 29878
2014 Jan 24
4
Booting Software RAID
I installed Centos 6.x 64 bit with the minimal ISO and used two disks in RAID 1 array. Filesystem Size Used Avail Use% Mounted on /dev/md2 97G 918M 91G 1% / tmpfs 16G 0 16G 0% /dev/shm /dev/md1 485M 54M 407M 12% /boot /dev/md3 3.4T 198M 3.2T 1% /vz Personalities : [raid1] md1 : active raid1 sda1[0] sdb1[1] 511936 blocks super 1.0
2007 Sep 04
4
RAID + LVM Addition to CentOS 5 Install
Hi All, I have what I believe to be a pretty basic LVM & RAID setup on my CentOS 5 machine: Raid Partitions: /dev/sda1,sdb1 /dev/sda2,sdb2 /dev/sda3,sdb3 During the install I created a RAID 1 volume md0 out of sda1,sdb1 for the boot partition and then added sda2,sdb2 to a separate RAID 1 volume as well (md1). I then setup md1 as a LVM physical volume for volume group 'system'. I
2011 Sep 07
1
boot problem after disk change on raid1
Hello, I have two disks sda and sdb. One of the was broken so I have changed the broken disk with a working one. I started the server in rescue mode, and created the partional table, and added all the partitions to the software raid. *I have added the partitions to the RAID, and reboot.* # mdadm /dev/md0 --add /dev/sdb1 # mdadm /dev/md1 --add /dev/sdb2 # mdadm /dev/md2 --add /dev/sdb3 # mdadm
2011 Mar 20
2
task md1_resync:9770 blocked for more than 120 seconds and OOM errors
Hello, yesterday night I had a problem with my server located at a hoster (strato.de). I couldn't ssh to it and over the remote serial console I saw "out of memory" errors (sorry, don't have the text). Then I had reinstall CentOS 5.5/64 bit + all my setup (2h work), because I have a contract with a social network and they will shut down my little card game if it is not
2010 Jan 05
4
Software RAID1 Disk I/O
I just installed CentOS 5.4 64 bit release on a 1.9ghz CPU with 8gB of RAM. It has 2 Western Digital 1.5TB SATA2 drives in RAID1. [root at server ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/md2 1.4T 1.4G 1.3T 1% / /dev/md0 99M 19M 76M 20% /boot tmpfs 4.0G 0 4.0G 0% /dev/shm [root at server ~]# Its barebones
2019 Apr 03
2
Kickstart putting /boot on sda2 (anaconda partition enumeration)?
Does anyone know how anaconda partitioning enumerates disk partitions when specified in kickstart? I quickly browsed through the anaconda installer source on github but didn't see the relevant bits. I'm using the centOS 6.10 anaconda installer. Somehow I am ending up with my swap partition on sda1, /boot on sda2, and root on sda3. for $REASONS I want /boot to be the partition #1 (sda1)
2007 Mar 06
1
blocks 256k chunks on RAID 1
Hi, I have a RAID 1 (using mdadm) on CentOS Linux and in /proc/mdstat I see this: md7 : active raid1 sda2[0] sdb2[1] 26627648 blocks [2/2] [UU] [-->> it's OK] md1 : active raid1 sdb3[1] sda3[0] 4192896 blocks [2/2] [UU] [-->> it's OK] md2 : active raid1 sda5[0] sdb5[1] 4192832 blocks [2/2] [UU] [-->> it's OK] md3 : active raid1 sdb6[1] sda6[0] 4192832 blocks [2/2]
2015 Feb 18
0
CentOS 7: software RAID 5 array with 4 disks and no spares?
Hi Niki, md127 apparently only uses 81.95GB per disk. Maybe one of the partitions has the wrong size. What's the output of lsblk? Regards Michael ----- Urspr?ngliche Mail ----- Von: "Niki Kovacs" <info at microlinux.fr> An: "CentOS mailing list" <CentOS at centos.org> Gesendet: Mittwoch, 18. Februar 2015 08:09:13 Betreff: [CentOS] CentOS 7: software RAID 5
2010 Sep 19
2
GRUB woes: GRUB Hard Disk Error
Hello all, I am attempting to install CentOS 5.5 on a recently donated home server (was Windows). I've tried installing twice; first, separating /boot, swap, /, /var, /usr and /home into different partitions, the second time choosing default layout. Both times have resulted in "GRUB Hard Disk Error" upon boot. About the hard disk, it is a Seagate SATA drive. I do not have another
2015 Aug 05
0
CentOS 5 grub boot problem
On 8/5/2015 4:00 PM, Chris Murphy wrote: > - Ahh OK now I see why I was confused. The originally posted partition > map uses cylinders as units, not LBA. I missed that. Cylinder 1 is the > same as LBA 63. And that is sufficiently large for a GRUB legacy stage > 2. > > - OK this is screwy. Partitions 1 and 3 on both drives have the same > number of sectors, but partitions 2
2008 Nov 10
1
Autodetecing RAID members upon boot... need to update initrd?
Hello fellow CentOS'ers- I've got a system running CentOS 5.0. The motherboard has two onboard SATA ports with two drives attached. I installed the system on a RAID1 setup. However, I'd like to add a hotspare disk to the array. Since there are no additional SATA ports, I've installed an additional controller. After partitioning, the additional drive was easily and successfully
2015 Aug 05
2
CentOS 5 grub boot problem
- Ahh OK now I see why I was confused. The originally posted partition map uses cylinders as units, not LBA. I missed that. Cylinder 1 is the same as LBA 63. And that is sufficiently large for a GRUB legacy stage 2. - OK this is screwy. Partitions 1 and 3 on both drives have the same number of sectors, but partitions 2 differ: /dev/hde2 401,625 975,691,709 975,290,085 fd Linux