similar to: create raid /dev/md2

Displaying 20 results from an estimated 1000 matches similar to: "create raid /dev/md2"

2005 Nov 08
1
EXT3-fs error (device md2): ext3_journal_start_sb: Detected aborted journal...
Hi, I'm running a production server (Debian Sarge install) whose root filesystem (a software raid 1 array of 2 partitions of IDE drive) exhibited the following problem: Oct 28 06:00:06 server2 kernel: attempt to access beyond end of device Oct 28 06:00:06 server2 kernel: md2: rw=1, want=3050401328, limit=16353920 [...] a few of the above line snipped, want is different each time Oct 28
2007 Dec 07
1
CentOS 5.1 on intel DQ35JO , Q35 chipsed based board
Hi, i was trying to install CentOS 5.1 on new machine based on Intel Q35 desktop board, witch 2 SATA disks (configured as RAID 1) , 8GB RAM and Intel Core2 QUADCORE. The first problem was with install - it hangs before installer startup at ACPI messages. So i installed it with ACPI=off switch. The first problem is : Dec 7 09:54:01 vmhost1 kernel: BIOS bug, no explicit IRQ entries, using default
2011 Apr 21
1
domain not present after shutdown
I hope this is simple I have one kvm host that lists all my domains even after shut down. 'virsh list --all' shows machines in 'running' and 'shut off' states. If I copy those over to another machine and it only shows the virt machines in running state. vmhost2:~$ virsh list --all Id Name State ---------------------------------- 4 mybox4
2012 Nov 30
1
How to repeat replication
Dear all, I generate data under IRT mdel. I conducted 1000 replications. When I run, some replication was not fit with my model. So, replications were fit model were less than 1000 replication. If I want its run until 1000 replicaions. How shouldI write functin. Thank you, Kamontip [[alternative HTML version deleted]]
2003 Jun 09
1
unable to read superblock ?
Hi, I have a machine setup as software-raid + ext3 + redhat , but this morning, it seems the file system crash. here is the booting message, anyone has any idea how to fix it. Thanks! Donghui ---------------------------------------------------------------------------- ------ md: adding sda1 md: created md2 md: using <sdb1> <sda1> md: md2: raid array is not clean -- starting
2011 May 04
2
Cannot resize btrfs volume
Hello, I added a new disk into our RAID5 array, it looks like this: md2 : active raid5 sdd4[3] sde4[4] sda4[0] sdc4[2] sdb4[1] 3767274240 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU] # btrfs fi sh Label: none uuid: 5534d2e7-be31-49c7-8ab7-90c5ab8afe18 Total devices 1 FS bytes used 2.24TB devid 3 size 2.63TB used 2.63TB path /dev/md2 # mount ... /dev/md2 on /home type btrfs
2010 May 28
2
permanently add md device
Hi All Currently i'm setting up a 5.4 server and try to create a 3rd raid device, when i run: $mdadm --create /dev/md2 -v --raid-devices=15 --chunk=32 --level=raid6 /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/sdk /dev/sdl /dev/sdm /dev/sdn /dev/sdo /dev/sdp /dev/sdq the device file "md2" is created and the raid is being configured. but somehow
2009 May 08
3
Software RAID resync
I have configured 2x 500G sata HDD as Software RAID1 with three partitions md0,md1 and md2 with md2 as 400+ gigs Now it is almost 36 hours the status is cat /proc/mdstat Personalities : [raid1] md0 : active raid1 hdb1[1] hda1[0] 104320 blocks [2/2] [UU] resync=DELAYED md1 : active raid1 hdb2[1] hda2[0] 4096448 blocks [2/2] [UU] resync=DELAYED md2 : active raid1
2007 Aug 02
1
"htree_dirblock_to_tree: bad entry in directory" error
Hi. I woke up this morning to find a ton of waiting emails complaining that some cron jobs on my system couldn't run because one of my filesystems (ext3 on software RAID 1) was suddenly mounted read-only. Always nice when you're away from the server due to travel. ;^> I investigated in the logs and found: 2007-08-02 04:02:25 kern.crit www kernel: EXT3-fs error (device md2):
2010 Aug 18
3
Wrong disk size problem.
Hi,we have Centos 5.4 server and according to me we have strange problem. Disk size and other indormation like below.Normally,md2 partition should have 46GB free disk size but available value is zero.Why it show zero ? If you help me,I will be happy. df -h Filesystem Size Used Avail Use% Mounted on /dev/md1 19G 2.1G 16G 12% / /dev/md2 880G 834G 0
2006 Aug 10
3
MD raid tools ... did i missed something?
Hi I have a degraded array /dev/md2 ===================================================================== $ mdadm -D /dev/md2 /dev/md2: Version : 00.90.01 Creation Time : Thu Oct 6 20:31:57 2005 Raid Level : raid5 Array Size : 221953536 (211.67 GiB 227.28 GB) Device Size : 110976768 (105.84 GiB 113.64 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 2
2007 Oct 17
2
Hosed my software RAID/LVM setup somehow
CentOS 5, original kernel (xen and normal) and everything, Linux RAID 1. I rebooted one of my machines after doing some changes to RAID/LVM and now the two RAID partitions that I made changes to are "gone". I cannot boot into the system. On bootup it tells me that the devices md2 and md3 are busy or mounted and drops me to the repair shell. When I run fs check manually it just tells
2010 Dec 29
1
list files on a device
Hello, After another power loss (I am fortune, am I not?) I have the following situation : Label: none uuid: ac155851-0e31-4aed-9ba4-ee712506368a Total devices 3 FS bytes used 1.02TB devid 1 size 931.51GB used 70.00GB path /dev/sdd1 devid 3 size 1.79TB used 66.52GB path /dev/md2 devid 2 size 914.70GB used 914.50GB path /dev/sda4 btrfs device delete does not work on both md2 and
2012 Jun 07
1
mdadm: failed to write superblock to
Hello, i have a little problem. Our server has an broken RAID. # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sda1[2](F) sdb1[1] 2096064 blocks [2/1] [_U] md2 : active raid1 sda3[2](F) sdb3[1] 1462516672 blocks [2/1] [_U] md1 : active raid1 sda2[0] sdb2[1] 524224 blocks [2/2] [UU] unused devices: <none> I have remove the partition: # mdadm --remove
2011 Dec 07
7
FS won't mount, open_ctree failed, Assertion !(path->slots[0] == 0) failed
So I''m having a bit of trouble with one of my btrfs filesystems. It isn''t mounting after a power failure. I can''t get restore or btrfsck to run, even on backup supers. I''ve pasted some output below. My btrfs-progs below come from git this morning, running on linux 3.2.0. # mount /dev/md2 /media/test/ mount: wrong fs type, bad option, bad superblock on
2014 Mar 31
2
Centos and Selinux issue
Hi list, I'm new to Centos and I've a very small knowledge of selinux use. I can disable it, but I prefer take it on for study. I've a second mirrored device that I use for file sharing. This is the scenario: /dev/md2 mounted on /mnt/data To make samba working I must set the file context to the path at samba_share_t on /mnt/data. After this samba works. Now I'm setting up
2011 May 21
1
Behavior difference in mbox versus Maildir listing
Hi, My MUA (claws-mail) is having a hard time listing directories for a mail account with two namespaces, one of which using mbox and the other one using maildir to store mails. Let's call the namespaces "#mbox." and "#maildir." and have "." as separator. Assume I have the following folder hierarchy: $namespace $namespace folder $namespace folder folder1a
2008 Mar 23
4
md raid1 - no speed improvement
Hi, I have two 320 GB SATA disks (/dev/sda, /dev/sdb) in a server running CentOS release 5. They both have three partitions setup as RAID1 using md (boot, swap, and an LVM data partition). # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sdb1[1] sda1[0] 104320 blocks [2/2] [UU] md1 : active raid1 sdb2[1] sda2[0] 4192896 blocks [2/2] [UU] md2 : active raid1 sdb3[1]
2006 Mar 02
3
Advice on setting up Raid and LVM
Hi all, I'm setting up Centos4.2 on 2x80GB SATA drives. The partition scheme is like this: /boot = 300MB / = 9.2GB /home = 70GB swap = 500MB The RAID is RAID 1. md0 = 300MB = /boot md1 = 9.2GB = LVM md2 = 70GB = LVM md3 = 500MB = LVM Now, the confusing part is: 1. When creating VolGroup00, should I include all PV (md1, md2, md3)? Then create the LV. 2. When setting up RAID 1, should I
2019 Feb 25
7
Problem with mdadm, raid1 and automatically adds any disk to raid
Hi. CENTOS 7.6.1810, fresh install - use this as a base to create/upgrade new/old machines. I was trying to setup two disks as a RAID1 array, using these lines mdadm --create --verbose /dev/md0 --level=0 --raid-devices=2 /dev/sdb1 /dev/sdc1 mdadm --create --verbose /dev/md1 --level=0 --raid-devices=2 /dev/sdb2 /dev/sdc2 mdadm --create --verbose /dev/md2 --level=0 --raid-devices=2