similar to: md raid1 - no speed improvement

Displaying 20 results from an estimated 9000 matches similar to: "md raid1 - no speed improvement"

2010 Jan 05
4
Software RAID1 Disk I/O
I just installed CentOS 5.4 64 bit release on a 1.9ghz CPU with 8gB of RAM. It has 2 Western Digital 1.5TB SATA2 drives in RAID1. [root at server ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/md2 1.4T 1.4G 1.3T 1% / /dev/md0 99M 19M 76M 20% /boot tmpfs 4.0G 0 4.0G 0% /dev/shm [root at server ~]# Its barebones
2014 Feb 07
3
Software RAID1 Failure Help
I am running software RAID1 on a somewhat critical server. Today I noticed one drive is giving errors. Good thing I had RAID. I planned on upgrading this server in next month or so. Just wandering if there was an easy way to fix this to avoid rushing the upgrade? Having a single drive is slowing down reads as well, I think. Thanks. Feb 7 15:28:28 server smartd[2980]: Device: /dev/sdb
2011 Sep 07
1
boot problem after disk change on raid1
Hello, I have two disks sda and sdb. One of the was broken so I have changed the broken disk with a working one. I started the server in rescue mode, and created the partional table, and added all the partitions to the software raid. *I have added the partitions to the RAID, and reboot.* # mdadm /dev/md0 --add /dev/sdb1 # mdadm /dev/md1 --add /dev/sdb2 # mdadm /dev/md2 --add /dev/sdb3 # mdadm
2012 Jun 07
1
mdadm: failed to write superblock to
Hello, i have a little problem. Our server has an broken RAID. # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sda1[2](F) sdb1[1] 2096064 blocks [2/1] [_U] md2 : active raid1 sda3[2](F) sdb3[1] 1462516672 blocks [2/1] [_U] md1 : active raid1 sda2[0] sdb2[1] 524224 blocks [2/2] [UU] unused devices: <none> I have remove the partition: # mdadm --remove
2013 Sep 15
1
grub command line
Hello Everyone I have a remote CentOS 6.4 server (with KVM access), when I received the server it was running with LVM on single disk (sda) I managed to remove LVM and install raid 1 in sda and sdb disks the mirroring is working fine, my only issue now is that everytime I reboot the server I got the grub command line and I have manually boot using comand grub> configfile
2010 Jul 23
5
install on raid1
Hi All, I'm currently trying to install centos 5.4 x86-64bit on a raid 1, so if one the 2 disks fails the server will still be available. i installed grub on /dev/sda using the advanced grub configuration option during the install. after the install is done i boot in linux rescue mode , chroot the filesystem and copy grub to both drives using: grub>root (hd0,0) grub>setup (hd0)
2012 Jan 13
5
Can't resize second device in RAID1
Hi, the situation: Label: ''RootFS''  uuid: c87975a0-a575-405e-9890-d3f7f25bbd96     Total devices 2 FS bytes used 284.98GB     devid    2 size 311.82GB used 286.51GB path /dev/sdb3     devid    1 size 897.76GB used 286.51GB path /dev/sda3 RootFS created when sda3 was 897.76GB and sdb3 311.82GB. I have now freed other space on sdb. So I deleted sdb3 and recreated it occupying all
2006 Apr 02
2
raid setup
Hi, I have 2 identical xSeries 346 with 2 identical IBM 72GB scsi drive. What i did is install the centos 4.2 serverCD on the first IBM and set the HDD to raid1 and raid0 for swap. Now what i did is get the 2nd HDD in the 1st Server swap it with the 1st HDD in the 2nd Server and rebuild the Raids. The 1st server rebuild the array fine. My problem is the Second server, after rebuilding it and
2016 Mar 12
4
C7 + UEFI + GPT + RAID1
Hi list, I'm new with UEFI and GPT. For several years I've used MBR partition table. I've installed my system on software raid1 (mdadm) using md0(sda1,sdb1) for swap, md1(sda2, sdb2) for /, md2 (sda3,sdb3) for /home. From several how-to concerning raid1 installation, I must put each partition on a different md devices. I've asked times ago if it's more correct create the
2015 Feb 18
0
CentOS 7: software RAID 5 array with 4 disks and no spares?
Hi Niki, md127 apparently only uses 81.95GB per disk. Maybe one of the partitions has the wrong size. What's the output of lsblk? Regards Michael ----- Urspr?ngliche Mail ----- Von: "Niki Kovacs" <info at microlinux.fr> An: "CentOS mailing list" <CentOS at centos.org> Gesendet: Mittwoch, 18. Februar 2015 08:09:13 Betreff: [CentOS] CentOS 7: software RAID 5
2014 Dec 12
0
Centos 7 how to make second disk of RAID1 bootable
Greetings - Ok, I have my CentOS 7 KVM host system installed and I want to be able to boot the system from either installed drive if one of them fails. My objective is to have the following layout for the two 3 TB disks. sda1 /boot/efi sda2 /boot sda3 RAID1 with sdb3 sdb1 /boot/efi sdb2 /boot sdb3 RAID1 with sda3 The system is installed and boots from sda[1,2] and md127 (sda3 and
2011 Mar 20
2
task md1_resync:9770 blocked for more than 120 seconds and OOM errors
Hello, yesterday night I had a problem with my server located at a hoster (strato.de). I couldn't ssh to it and over the remote serial console I saw "out of memory" errors (sorry, don't have the text). Then I had reinstall CentOS 5.5/64 bit + all my setup (2h work), because I have a contract with a social network and they will shut down my little card game if it is not
2014 Jan 24
4
Booting Software RAID
I installed Centos 6.x 64 bit with the minimal ISO and used two disks in RAID 1 array. Filesystem Size Used Avail Use% Mounted on /dev/md2 97G 918M 91G 1% / tmpfs 16G 0 16G 0% /dev/shm /dev/md1 485M 54M 407M 12% /boot /dev/md3 3.4T 198M 3.2T 1% /vz Personalities : [raid1] md1 : active raid1 sda1[0] sdb1[1] 511936 blocks super 1.0
2007 Sep 04
4
RAID + LVM Addition to CentOS 5 Install
Hi All, I have what I believe to be a pretty basic LVM & RAID setup on my CentOS 5 machine: Raid Partitions: /dev/sda1,sdb1 /dev/sda2,sdb2 /dev/sda3,sdb3 During the install I created a RAID 1 volume md0 out of sda1,sdb1 for the boot partition and then added sda2,sdb2 to a separate RAID 1 volume as well (md1). I then setup md1 as a LVM physical volume for volume group 'system'. I
2019 Feb 25
0
Problem with mdadm, raid1 and automatically adds any disk to raid
In article <20190225050144.GA5984 at button.barrett.com.au>, Jobst Schmalenbach <jobst at barrett.com.au> wrote: > Hi. > > CENTOS 7.6.1810, fresh install - use this as a base to create/upgrade new/old machines. > > I was trying to setup two disks as a RAID1 array, using these lines > > mdadm --create --verbose /dev/md0 --level=0 --raid-devices=2 /dev/sdb1
2007 Mar 06
1
blocks 256k chunks on RAID 1
Hi, I have a RAID 1 (using mdadm) on CentOS Linux and in /proc/mdstat I see this: md7 : active raid1 sda2[0] sdb2[1] 26627648 blocks [2/2] [UU] [-->> it's OK] md1 : active raid1 sdb3[1] sda3[0] 4192896 blocks [2/2] [UU] [-->> it's OK] md2 : active raid1 sda5[0] sdb5[1] 4192832 blocks [2/2] [UU] [-->> it's OK] md3 : active raid1 sdb6[1] sda6[0] 4192832 blocks [2/2]
2013 Mar 03
4
Strange behavior from software RAID
Somewhere, mdadm is cacheing information. Here is my /etc/mdadm.conf file: more /etc/mdadm.conf # mdadm.conf written out by anaconda DEVICE partitions MAILADDR root ARRAY /dev/md0 level=raid1 num-devices=4 metadata=0.90 UUID=55ff58b2:0abb5bad:42911890:5950dfce ARRAY /dev/md1 level=raid1 num-devices=2 metadata=0.90 UUID=315eaf5c:776c85bd:5fa8189c:68a99382 ARRAY /dev/md2 level=raid1 num-devices=2
2010 Jun 22
1
The Host cpu(s) in this machine do not have support for full virtualization
Hi, I'm geting this message from virt-manager. Server: SuperMicro X8SIL with INTEL X3440 Here is my logs /var/log/dmesg: Bootdata ok (command line is ro root=/dev/md2) Linux version 2.6.18-194.3.1.el5xen (mockbuild at builder10.centos.org) (gcc version 4.1.2 20080704 (Red Hat 4.1.2-48)) #1 SMP Thu May 13 13:49:53 EDT 2010 BIOS-provided physical RAM map: Xen: 0000000000000000 -
2015 Feb 18
5
CentOS 7: software RAID 5 array with 4 disks and no spares?
Hi, I just replaced Slackware64 14.1 running on my office's HP Proliant Microserver with a fresh installation of CentOS 7. The server has 4 x 250 GB disks. Every disk is configured like this : * 200 MB /dev/sdX1 for /boot * 4 GB /dev/sdX2 for swap * 248 GB /dev/sdX3 for / There are supposed to be no spare devices. /boot and swap are all supposed to be assembled in RAID level 1 across
2019 Feb 25
0
Problem with mdadm, raid1 and automatically adds any disk to raid
> Hi. > > CENTOS 7.6.1810, fresh install - use this as a base to create/upgrade > new/old machines. > > I was trying to setup two disks as a RAID1 array, using these lines > > mdadm --create --verbose /dev/md0 --level=0 --raid-devices=2 /dev/sdb1 > /dev/sdc1 > mdadm --create --verbose /dev/md1 --level=0 --raid-devices=2 /dev/sdb2 > /dev/sdc2 > mdadm