similar to: Help interpreting RAID1 space allocation

Displaying 20 results from an estimated 20000 matches similar to: "Help interpreting RAID1 space allocation"

2012 Oct 26
4
Can't replace a faulty disk of raid1
Hello, I had a raid1 btrfs (540GB) on vanilla 3.6.3, a disk failed, and removed it at power off, plugged in a new one, partitioned it (to 110GB, by error), and added it to btrfs. I tried to remove the missing device, and it said "Input/output error" after a while. Next attempts simply gave "Invalid argument". I repartitioned, rebooted the system, and made the partition grow:
2012 Apr 17
3
Btrfs in degraded mode
Hello, I have created a btrfs filesystem with RAID1 setup having 2 disks. Everything works fine but when I try to umount the device and remount it in degraded mode, the data still goes into both the disk. ideally in degraded mode only one disk show disk activity and not the failed ones. System Config: Base OS: Slackware kernel: linux 3.3.2 "sar -pd 2 10" shows me that the data is
2013 Dec 07
8
Can't remove empty directory after kernel panic, no errors in dmesg
Hi List, so first the basics. I''m running Arch Linux with 3.13-rc2, btrfs-progs 0.20rc1.3-2 from the repo and I''m using a SSD. So I was having kernel panics with my USB 3.0 Gigabit card and was trying to get a panic output. These panics are intermittent and most often happen while using Chromium. Anyway so my system paniced while I was in Chromium. After the reboot Chromium
2013 Oct 06
5
btrfs device delete problem
Hi, I''m getting an error when trying to delete a device from a raid1 (data and metadata mirrored). > btrfs filesystem show failed to read /dev/sr0 Label: none uuid: 78b5162b-489e-4de1-a989-a47b91adef50 Total devices 2 FS bytes used 107.64GB devid 2 size 149.05GB used 109.01GB path /dev/sdh1 devid 1 size 156.81GB used 109.03GB path /dev/sdb6 Btrfs v0.20-rc1 >
2011 Aug 14
3
cant mount degraded (it worked in kernel 2.6.38.8)
# uname -a Linux dhcppc1 3.0.1-xxxx-std-ipv6-64 #1 SMP Sun Aug 14 17:06:21 CEST 2011 x86_64 x86_64 x86_64 GNU/Linux mkdir test5 cd test5 dd if=/dev/null of=img5 bs=1 seek=2G dd if=/dev/null of=img6 bs=1 seek=2G losetup /dev/loop2 img5 losetup /dev/loop3 img6 mkfs.btrfs -d raid1 -m raid1 /dev/loop2 /dev/loop3 btrfs device scan btrfs filesystem show Label: none uuid:
2012 Oct 25
46
[RFC] New attempt to a better "btrfs fi df"
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi all, this is a new attempt to improve the output of the command "btrfs fi df". The previous attempt received a good reception. However there was no a general consensus about the wording. Moreover I still didn''t understand how btrfs was using the disks. A my first attempt was to develop a new command which shows how the disks
2011 Jul 12
1
after mounting with -o degraded: ioctl: LOOP_CLR_FD: Device or resource busy
dd if=/dev/null of=img5 bs=1 seek=2G dd if=/dev/null of=img6 bs=1 seek=2G mkfs.btrfs -d raid1 -m raid1 img5 img6 losetup /dev/loop4 img5 losetup /dev/loop5 img6 btrfs device scan mount -t btrfs /dev/loop4 dir umount dir losetup -d /dev/loop5 mount -t btrfs -o degraded /dev/loop4 dir umount dir losetup -d /dev/loop4 ioctl: LOOP_CLR_FD: Device or resource busy mkfs.ext3 /dev/loop4 mke2fs 1.39
2011 Jan 12
1
Filesystem creation in "degraded mode"
I''ve had a go at determining exactly what happens when you create a filesystem without enough devices to meet the requested replication strategy: # mkfs.btrfs -m raid1 -d raid1 /dev/vdb # mount /dev/vdb /mnt # btrfs fi df /mnt Data: total=8.00MB, used=0.00 System, DUP: total=8.00MB, used=4.00KB System: total=4.00MB, used=0.00 Metadata, DUP: total=153.56MB, used=24.00KB Metadata:
2013 Jul 27
5
Mount multiple-device-filesystem by UUID
Hello, As stated in the -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
2012 Jan 22
1
Trying to mount RAID1 degraded with removed disk -> open_ctree failed
Hi, I have setup a RAID1 using 3 devices (500G each) on separate disks. After removing one disk physically the filesystem cannot be mounted in degraded nor in recovery mode. When putting the disk back in the filesystem can be mounted without errors. I did a cold-swap (powercycle after removal/insertion of the disk). Here are the details: - latest kernel 3.2.1 and btrfs-tools on xubuntu
2012 Dec 17
5
Feeback on RAID1 feature of Btrfs
Hello, I''m testing Btrfs RAID1 feature on 3 disks of ~10GB. Last one is not exactly 10GB (would be too easy). About the test machine, it''s a kvm vm running an up-to-date archlinux with linux 3.7 and btrfs-progs 0.19.20121005. #uname -a Linux seblu-btrfs-1 3.7.0-1-ARCH #1 SMP PREEMPT Tue Dec 11 15:05:50 CET 2012 x86_64 GNU/Linux Filesystem was created with : # mkfs.btrfs -L
2010 Jan 22
1
R: Re: mount command not honoring rootflags passed
Hi Thomas, >----Messaggio originale---- >Da: thomas@kupper.org >Data: 22/01/2010 10.48 >A: "Leszek Ciesielski"<skolima@gmail.com> >Cc: "linux-btrfs"<linux-btrfs@vger.kernel.org> >Ogg: Re: mount command not honoring rootflags passed > > >On 22 Jan 2010, at 10:40, Leszek Ciesielski wrote: > >> On Fri, Jan 22, 2010 at 10:30 AM,
2012 Jan 17
8
[RFC][PATCH 1/2] Btrfs: try to allocate new chunks with degenerated profile
If there is no free space, the free space allocator will try to get space from the block group with the degenerated profile. For example, if there is no free space in the RAID1 block groups, the allocator will try to allocate space from the DUP block groups. And besides that, the space reservation has the similar behaviour: if there is no enough space in the space cache to reserve, it will reserve
2010 Nov 22
9
btrfs problems and fedora 14
I thought I would try btrfs on a new installation of f14. yes, I know its experimental but stable so it seemed to be a good time to try it. I am not sure if I have missed something out of all my searching but am I correct in thinking that currently: I. it is not possible to boot from a snapshot of the operating system and, in particular, the yum snapshots cannot be used for
2011 Aug 08
7
“bio too big” regression and silent data corruption in 3.0
tl;dr version: 3.0 produces “bio too big” dmesg entries and silently corrupts data in “meta-raid1/data-single” configurations on disks with different max_hw_sectors, where 2.6.38 worked fine. tl;dr side-issue: on-line removal of partitions holding “single” data attempts to create raid0 (rather than single) block groups. If it can''t get enough room for raid0 over all remaining disks, it
2013 Jun 03
3
csum failed during rebalance
Hi, I added a new drive to an existing RAID 0 array. Every attempt to rebalance the array fails: # btrfs filesystem balance /share/bd8 ERROR: error during balancing ''/share/bd8'' - Input/output error # dmesg | tail btrfs: found 1 extents btrfs: relocating block group 10752513540096 flags 1 btrfs: found 5 extents btrfs: found 5 extents btrfs: relocating block group 10751439798272
2013 Oct 07
2
Some questions after devices addition to existing raid 1 btrfs filesystem
Hi, I have added 2x2Tb to my existing 2x2Tb raid 1 btrfs filesystem and then ran a balance: # btrfs filesystem show Total devices 4 FS bytes used 1.74TB devid 3 size 1.82TB used 0.00 path /dev/sdd devid 4 size 1.82TB used 0.00 path /dev/sde devid 2 size 1.82TB used 1.75TB path /dev/sdc devid 1 size 1.82TB used 1.75TB path /dev/sdb # btrfs
2009 Nov 19
10
Unable to mount loopback devices in RAID mode
Hi! I recently tried to mount a filesystem in RAID1 mode using loopback devices. I followed the instructions at [1]. Here''s exactly what I''ve done: $ dd if=/dev/zero of=raid1_0.img bs=1M count=500 $ dd if=/dev/zero of=raid1_1.img bs=1M count=500 $ mkfs.btrfs -m raid1 -d raid1 raid1_0.img raid1_1.img $ losetup /dev/loop0 raid1_0.img $ losetup /dev/loop1 raid1_1.img $ mount -t
2013 May 13
7
Remove a materially failed device from a Btrfs "single-raid" using partitions
Hello, I am on Ubuntu Server 13.04 with Linux 3.8. I''ve created a "single-raid" using /dev/sd{a,b,c,d}{1,3}. One of my hard drives has failed, I mean it''s materially dead. :~$ sudo btrfs filesystem show Label: none uuid: 40886f51-8c9b-4be1-8721-83bf5653d2a0 Total devices 5 FS bytes used 226.90GB devid 4 size 37.27GB used 31.01GB path /dev/sdd1
2013 Aug 04
2
Unable to unmount filesystem (bug in kernel reported in kern.log)
I tried to unmount a btrfs filesystem located in a external usb hard drive. This belonged to a raid1 data and metadata filesystem mounted in degraded mode. Unfortunately, I couldn''t save the image of filesystem but I could see this error in kern.log: Aug 4 02:23:55 rohan kernel: [ 3747.840027] usb 1-3: new high-speed USB device number 8 using ehci_hcd Aug 4 02:23:55 rohan kernel: [