Displaying 20 results from an estimated 10000 matches similar to: "Software RAID1 with CentOS-6.2"
2010 Oct 19
3
more software raid questions
hi all!
back in Aug several of you assisted me in solving a problem where one
of my drives had dropped out of (or been kicked out of) the raid1 array.
something vaguely similar appears to have happened just a few mins ago,
upon rebooting after a small update. I received four emails like this,
one for /dev/md0, one for /dev/md1, one for /dev/md125 and one for
/dev/md126:
Subject: DegradedArray
2006 Mar 14
2
Help. Failed event on md1
Hi all,
This morning I received this notification from mdadm:
This is an automatically generated mail message from mdadm
running on server-mail.mydomain.kom
A Fail event had been detected on md device /dev/md1.
Faithfully yours, etc.
In /proc/mdstat I see this:
Personalities : [raid1]
md1 : active raid1 sdb2[2](F) sda2[0]
77842880 blocks [2/1] [U_]
md0 : active raid1 sdb1[1] sda1[0]
2019 Apr 09
2
Kernel panic after removing SW RAID1 partitions, setting up ZFS.
System is CentOS 6 all up to date, previously had two drives in MD RAID
configuration.
md0: sda1/sdb1, 20 GB, OS / Partition
md1: sda2/sdb2, 1 TB, data mounted as /home
Installed kmod ZFS via yum, reboot, zpool works fine. Backed up the /home data
2x, then stopped the sd[ab]2 partition with:
mdadm --stop /dev/md1;
mdadm --zero-superblock /dev/sd[ab]1;
Removed /home in /etc/fstab. Used
2019 Feb 25
7
Problem with mdadm, raid1 and automatically adds any disk to raid
Hi.
CENTOS 7.6.1810, fresh install - use this as a base to create/upgrade new/old machines.
I was trying to setup two disks as a RAID1 array, using these lines
mdadm --create --verbose /dev/md0 --level=0 --raid-devices=2 /dev/sdb1 /dev/sdc1
mdadm --create --verbose /dev/md1 --level=0 --raid-devices=2 /dev/sdb2 /dev/sdc2
mdadm --create --verbose /dev/md2 --level=0 --raid-devices=2
2016 Mar 12
4
C7 + UEFI + GPT + RAID1
Hi list,
I'm new with UEFI and GPT.
For several years I've used MBR partition table. I've installed my
system on software raid1 (mdadm) using md0(sda1,sdb1) for swap,
md1(sda2, sdb2) for /, md2 (sda3,sdb3) for /home. From several how-to
concerning raid1 installation, I must put each partition on a different
md devices. I've asked times ago if it's more correct create the
2007 Sep 25
2
mdadm problem.
So I'm trying to RAID-1 this system which has two identical disks
installed in it, and it isn't working for some reason.
I started by doing a CentOS-4 install on /dev/sda1 as root, and with
/dev/sda2 as my swap.
I finish the install, yum update, and then I want to make the mirrors.
I copy the partition table from one disk to the other:
# sfdisk -d /dev/sda | sfdisk /dev/sdb
I create
2009 Jul 02
4
Upgrading drives in raid 1
I think I have solved my issue and would like some input from anyone who has
done this for pitfalls, errors, or if I am just wrong.
Centos 5.x, software raid, 250gb drives.
2 drives in mirror, one spare. All same size.
2 devices in the mirror, one boot (about 100MB), one that fills the rest of
disk and contains LVM partitions.
I was thinking of taking out the spare and adding a 500gb drive.
I
2014 Feb 07
3
Software RAID1 Failure Help
I am running software RAID1 on a somewhat critical server. Today I
noticed one drive is giving errors. Good thing I had RAID. I planned
on upgrading this server in next month or so. Just wandering if there
was an easy way to fix this to avoid rushing the upgrade? Having a
single drive is slowing down reads as well, I think.
Thanks.
Feb 7 15:28:28 server smartd[2980]: Device: /dev/sdb
2008 Apr 17
2
Question about RAID 5 array rebuild with mdadm
I'm using Centos 4.5 right now, and I had a RAID 5 array stop because
two drives became unavailable. After adjusting the cables on several
occasions and shutting down and restarting, I was able to see the
drives again. This is when I snatched defeat from the jaws of
victory. Please, someone with vast knowledge of how RAID 5 with mdadm
works, tell me if I have any chance at all
2022 Apr 24
3
Installing mdadm and C7 on new computer
On 04/23/2022 09:19 PM, H wrote:
> On 04/19/2022 09:57 AM, Roberto Ragusa wrote:
>> On 4/18/22 1:27 PM, H wrote:
>>> I have a new computer with 2 x 2TB SSDs where I wanted to install C7 and use mdadm for RAID1 configuration and encrypting the /home partition. On the net I found https://tuxfixer.com/centos-7-installation-with-lvm-raid-1-mirroring/ which I adopted slightly with
2014 Dec 03
7
DegradedArray message
Received the following message in mail to root:
Message 257:
>From root at desk4.localdomain Tue Oct 28 07:25:37 2014
Return-Path: <root at desk4.localdomain>
X-Original-To: root
Delivered-To: root at desk4.localdomain
From: mdadm monitoring <root at desk4.localdomain>
To: root at desk4.localdomain
Subject: DegradedArray event on /dev/md0:desk4
Date: Tue, 28 Oct 2014 07:25:27
2012 Jun 07
1
mdadm: failed to write superblock to
Hello,
i have a little problem. Our server has an broken RAID.
# cat /proc/mdstat
Personalities : [raid1]
md0 : active raid1 sda1[2](F) sdb1[1]
2096064 blocks [2/1] [_U]
md2 : active raid1 sda3[2](F) sdb3[1]
1462516672 blocks [2/1] [_U]
md1 : active raid1 sda2[0] sdb2[1]
524224 blocks [2/2] [UU]
unused devices: <none>
I have remove the partition:
# mdadm --remove
2008 Dec 12
1
Upgrade to new drives in raid, larger
Hi all,
As part of my raid experience, I have yet to have to do this, but was
wondering how you guys would attempt it.
I have 3 drives in a raid 1, with one as a hot spare.
They are 250gb with all space used by two raid devices, 1 with boot, the
other with LVMs filling them up.
Now, lets say down the road I want to put in 500gb drives and replace
them....yikes.
I was thinking of taking out the
2016 Mar 01
10
Any experiences with newer WD Red drives?
Might be slightly OT as it isn't necessarily a CentOS related issue.
I've been using WD Reds as mdraid components which worked pretty well
for non-IOPS intensive workloads.
However, the latest C7 server I built, ran into problems with them on
on a Intel C236 board (SuperMicro X11SSH) with tons of "ata bus error
write fpdma queued". Googling on it threw up old suggestions to
2010 Jun 29
1
ZFS on Caviar Blue (Hard Drive Recommendations)
Hi list,
I googled around but couldn''t find anything on whether someone has
good or bad experiences with the Caviar *Blue* drives? I saw in the
archives Caviar Blacks are *not* recommended for ZFS arrays (excluding
apparently RE3 and RE4?). Specifically I''m looking to buy Western
Digital Caviar Blue WD10EALS 1TB drives [1]. Does anyone have any
experience with these drives?
If
2010 Jun 09
2
software raid - better management advice needed
Hi,
I've used mdadm for years now to manage software raids.
The task of using fdisk to first create partitions on a spare drive
sitting on a shelf (raid 0 were my 1st of 2 drives failed) is kind of
bugging me now.
After using fdisk to create the same partition layout on the new drive
as is on the existing drive and then using mdadm to finish every thing
up is a little tedious.
Any
2014 Dec 09
2
DegradedArray message
On Thu, 2014-12-04 at 16:46 -0800, Gordon Messmer wrote:
> On 12/04/2014 05:45 AM, David McGuffey wrote:
> In practice, however, there's a bunch of information you didn't provide,
> so some of those steps are wrong.
>
> I'm not sure what dm-0, dm-2 and dm-3 are, but they're indicated in your
> mdstat. I'm guessing that you made partitions, and then made
2013 Feb 04
3
Questions about software RAID, LVM.
I am planning to increase the disk space on my desktop system. It is
running CentOS 5.9 w/XEN. I have two 160Gig 2.5" laptop (2.5") SATA drives
in two slots of a 4-slot hot swap bay configured like this:
Disk /dev/sda: 160.0 GB, 160041885696 bytes
255 heads, 63 sectors/track, 19457 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot Start End
2019 Mar 12
1
CentOS 7 Installation Problems
I attempted to install CentOS 7 x86_64 on my machine that has the
following hardware:
Motherboard:????? ASRock X99 Taichi
BIOS:???????????? AMI v P1.40? 08/04/2016
CPU:????????????? Intel Core I7-5820K
RAM:????????????? 64 GB (8 x 8 GB DIMM)
Optical:????????? LG Blu Ray 25 G / 50 G burner
Storage:????????? 2 - 120GB?? PNY CS1311 SSD? 4 - 4 TB Western Digital
hard drives
????????????????? The 2
2015 Feb 09
2
Another Fedora decision
On 10 February 2015 at 10:08, Kahlil Hodgson
<kahlil.hodgson at dealmax.com.au> wrote:
> I think you can simply submit a bug report under fedora documentation.
Via bugzilla:
https://bugzilla.redhat.com/enter_bug.cgi?product=Fedora%20Documentation