similar to: unable to read superblock ?

Displaying 20 results from an estimated 1000 matches similar to: "unable to read superblock ?"

2012 Jun 07
1
mdadm: failed to write superblock to
Hello, i have a little problem. Our server has an broken RAID. # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sda1[2](F) sdb1[1] 2096064 blocks [2/1] [_U] md2 : active raid1 sda3[2](F) sdb3[1] 1462516672 blocks [2/1] [_U] md1 : active raid1 sda2[0] sdb2[1] 524224 blocks [2/2] [UU] unused devices: <none> I have remove the partition: # mdadm --remove
2011 May 04
2
Cannot resize btrfs volume
Hello, I added a new disk into our RAID5 array, it looks like this: md2 : active raid5 sdd4[3] sde4[4] sda4[0] sdc4[2] sdb4[1] 3767274240 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU] # btrfs fi sh Label: none uuid: 5534d2e7-be31-49c7-8ab7-90c5ab8afe18 Total devices 1 FS bytes used 2.24TB devid 3 size 2.63TB used 2.63TB path /dev/md2 # mount ... /dev/md2 on /home type btrfs
2010 May 28
2
permanently add md device
Hi All Currently i'm setting up a 5.4 server and try to create a 3rd raid device, when i run: $mdadm --create /dev/md2 -v --raid-devices=15 --chunk=32 --level=raid6 /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/sdk /dev/sdl /dev/sdm /dev/sdn /dev/sdo /dev/sdp /dev/sdq the device file "md2" is created and the raid is being configured. but somehow
2009 May 08
3
Software RAID resync
I have configured 2x 500G sata HDD as Software RAID1 with three partitions md0,md1 and md2 with md2 as 400+ gigs Now it is almost 36 hours the status is cat /proc/mdstat Personalities : [raid1] md0 : active raid1 hdb1[1] hda1[0] 104320 blocks [2/2] [UU] resync=DELAYED md1 : active raid1 hdb2[1] hda2[0] 4096448 blocks [2/2] [UU] resync=DELAYED md2 : active raid1
2007 Aug 02
1
"htree_dirblock_to_tree: bad entry in directory" error
Hi. I woke up this morning to find a ton of waiting emails complaining that some cron jobs on my system couldn't run because one of my filesystems (ext3 on software RAID 1) was suddenly mounted read-only. Always nice when you're away from the server due to travel. ;^> I investigated in the logs and found: 2007-08-02 04:02:25 kern.crit www kernel: EXT3-fs error (device md2):
2010 Aug 18
3
Wrong disk size problem.
Hi,we have Centos 5.4 server and according to me we have strange problem. Disk size and other indormation like below.Normally,md2 partition should have 46GB free disk size but available value is zero.Why it show zero ? If you help me,I will be happy. df -h Filesystem Size Used Avail Use% Mounted on /dev/md1 19G 2.1G 16G 12% / /dev/md2 880G 834G 0
2008 Oct 23
5
compare directories
Hello guys, I have two mirrors. I need to compare files and directories on both mirrors and as a result print list of those which are missing on mirror 2 What i did find /data > find.mirror1 find /data > find.mirror2 Now i need to get list of those directories which are missing in mirror1. Thanks in advance! David -------------- next part -------------- An HTML attachment was
2005 Nov 08
1
EXT3-fs error (device md2): ext3_journal_start_sb: Detected aborted journal...
Hi, I'm running a production server (Debian Sarge install) whose root filesystem (a software raid 1 array of 2 partitions of IDE drive) exhibited the following problem: Oct 28 06:00:06 server2 kernel: attempt to access beyond end of device Oct 28 06:00:06 server2 kernel: md2: rw=1, want=3050401328, limit=16353920 [...] a few of the above line snipped, want is different each time Oct 28
2006 Aug 10
3
MD raid tools ... did i missed something?
Hi I have a degraded array /dev/md2 ===================================================================== $ mdadm -D /dev/md2 /dev/md2: Version : 00.90.01 Creation Time : Thu Oct 6 20:31:57 2005 Raid Level : raid5 Array Size : 221953536 (211.67 GiB 227.28 GB) Device Size : 110976768 (105.84 GiB 113.64 GB) Raid Devices : 3 Total Devices : 2 Preferred Minor : 2
2007 Oct 17
2
Hosed my software RAID/LVM setup somehow
CentOS 5, original kernel (xen and normal) and everything, Linux RAID 1. I rebooted one of my machines after doing some changes to RAID/LVM and now the two RAID partitions that I made changes to are "gone". I cannot boot into the system. On bootup it tells me that the devices md2 and md3 are busy or mounted and drops me to the repair shell. When I run fs check manually it just tells
2010 Dec 29
1
list files on a device
Hello, After another power loss (I am fortune, am I not?) I have the following situation : Label: none uuid: ac155851-0e31-4aed-9ba4-ee712506368a Total devices 3 FS bytes used 1.02TB devid 1 size 931.51GB used 70.00GB path /dev/sdd1 devid 3 size 1.79TB used 66.52GB path /dev/md2 devid 2 size 914.70GB used 914.50GB path /dev/sda4 btrfs device delete does not work on both md2 and
2011 Dec 07
7
FS won't mount, open_ctree failed, Assertion !(path->slots[0] == 0) failed
So I''m having a bit of trouble with one of my btrfs filesystems. It isn''t mounting after a power failure. I can''t get restore or btrfsck to run, even on backup supers. I''ve pasted some output below. My btrfs-progs below come from git this morning, running on linux 3.2.0. # mount /dev/md2 /media/test/ mount: wrong fs type, bad option, bad superblock on
2014 Mar 31
2
Centos and Selinux issue
Hi list, I'm new to Centos and I've a very small knowledge of selinux use. I can disable it, but I prefer take it on for study. I've a second mirrored device that I use for file sharing. This is the scenario: /dev/md2 mounted on /mnt/data To make samba working I must set the file context to the path at samba_share_t on /mnt/data. After this samba works. Now I'm setting up
2011 May 21
1
Behavior difference in mbox versus Maildir listing
Hi, My MUA (claws-mail) is having a hard time listing directories for a mail account with two namespaces, one of which using mbox and the other one using maildir to store mails. Let's call the namespaces "#mbox." and "#maildir." and have "." as separator. Assume I have the following folder hierarchy: $namespace $namespace folder $namespace folder folder1a
2008 Mar 23
4
md raid1 - no speed improvement
Hi, I have two 320 GB SATA disks (/dev/sda, /dev/sdb) in a server running CentOS release 5. They both have three partitions setup as RAID1 using md (boot, swap, and an LVM data partition). # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 sdb1[1] sda1[0] 104320 blocks [2/2] [UU] md1 : active raid1 sdb2[1] sda2[0] 4192896 blocks [2/2] [UU] md2 : active raid1 sdb3[1]
2006 Mar 02
3
Advice on setting up Raid and LVM
Hi all, I'm setting up Centos4.2 on 2x80GB SATA drives. The partition scheme is like this: /boot = 300MB / = 9.2GB /home = 70GB swap = 500MB The RAID is RAID 1. md0 = 300MB = /boot md1 = 9.2GB = LVM md2 = 70GB = LVM md3 = 500MB = LVM Now, the confusing part is: 1. When creating VolGroup00, should I include all PV (md1, md2, md3)? Then create the LV. 2. When setting up RAID 1, should I
2008 Apr 18
1
create raid /dev/md2
Hi , currently i have 2 raid devices /dev/md0 and /dev/md1 , i have added 2 new disks, fdisked , created 2 primary partitions with type fd (linux raid autodetect) Now i want to create raid from them root at vmhost1 ~]# mdadm --create --verbose /dev/md2 --level=1 /dev/sdc1 /dev/sdd1 mdadm: error opening /dev/md2: No such file or directory will return that error, what shouldi do? Thanks!
2009 Feb 23
1
Interleave or not
Lets say you had 4 servers and you wanted to setup replicate and distribute. What methoid would be better: server sdb1 xen0 brick0 xen1 mirror0 xen2 brick1 xen3 mirror1 replicate block0 - brick0 mirror0 replicate block1 - brick1 mirror1 distribute unify - block0 block1 or server sdb1 sdb2 xen0 brick0 mirror3 xen1 brick1 mirror0 xen2 brick2 mirror1 xen3 brick3 mirror2 replicate block0 -
2019 Feb 25
7
Problem with mdadm, raid1 and automatically adds any disk to raid
Hi. CENTOS 7.6.1810, fresh install - use this as a base to create/upgrade new/old machines. I was trying to setup two disks as a RAID1 array, using these lines mdadm --create --verbose /dev/md0 --level=0 --raid-devices=2 /dev/sdb1 /dev/sdc1 mdadm --create --verbose /dev/md1 --level=0 --raid-devices=2 /dev/sdb2 /dev/sdc2 mdadm --create --verbose /dev/md2 --level=0 --raid-devices=2
2007 Sep 25
2
mdadm problem.
So I'm trying to RAID-1 this system which has two identical disks installed in it, and it isn't working for some reason. I started by doing a CentOS-4 install on /dev/sda1 as root, and with /dev/sda2 as my swap. I finish the install, yum update, and then I want to make the mirrors. I copy the partition table from one disk to the other: # sfdisk -d /dev/sda | sfdisk /dev/sdb I create