similar to: some feedbacks seen on btrfs

Displaying 20 results from an estimated 10000 matches similar to: "some feedbacks seen on btrfs"

2011 Apr 09
16
wrong values in "df" and "btrfs filesystem df"
Hallo, linux-btrfs, First I create an array of 2 disks with mkfs.btrfs -d raid0 -m raid1 /dev/sdb1 /dev/sdd1 and mount it at /srv/MM. Then I fill it with about 1,6 TByte. And then I add /dev/sde1 via btrfs device add /dev/sde1 /srv/MM btrfs filesystem balance /srv/MM (it run about 20 hours) Then I work on it, copy some new files, delete some old files - all works well. Only df
2013 Aug 16
3
4 vol raid5 segfault on device delete
I have a 4 device volume with raid5 - trying to remove one of the devices (plenty of free space) and I get an almost immediate segfault. Scrub shows no errors, repair show space cache invalid but nothing else (I remounted with clear cache to be safe). Lots of corrupt on bdev (for 3 out of 4 drives), but I have no file access issues that I know of. Thanks! Output below:
2009 Nov 19
10
Unable to mount loopback devices in RAID mode
Hi! I recently tried to mount a filesystem in RAID1 mode using loopback devices. I followed the instructions at [1]. Here''s exactly what I''ve done: $ dd if=/dev/zero of=raid1_0.img bs=1M count=500 $ dd if=/dev/zero of=raid1_1.img bs=1M count=500 $ mkfs.btrfs -m raid1 -d raid1 raid1_0.img raid1_1.img $ losetup /dev/loop0 raid1_0.img $ losetup /dev/loop1 raid1_1.img $ mount -t
2013 Apr 03
2
[bug] btrfs fi df doesn't show raid type after balance
Did something break.. ? we are not reporting raid type after balance. ----------- # btrfs fi df /btrfs Data, RAID0: total=2.00GB, used=2.03MB Data: total=8.00MB, used=0.00 System, RAID0: total=16.00MB, used=4.00KB System: total=4.00MB, used=0.00 Metadata, RAID0: total=2.00GB, used=216.00KB Metadata: total=8.00MB, used=4.00KB # btrfs bal /btrfs Done, had to relocate 5 out of 5 chunks # btrfs fi
2013 Oct 14
1
Many questions from a potential btrfs user
Hi. I am seriously considering employing btrfs on my systems, particularly due to some space-saving features that it has (namely, deduplication and compression). In fact, I was (a few moments ago) trying to back up some of my systems to a 2TB HD that has an ext4 filesystem and, in the middle of the last one, I got the error message that the backup HD was full. Given that what I backup there are
2013 Oct 06
5
btrfs device delete problem
Hi, I''m getting an error when trying to delete a device from a raid1 (data and metadata mirrored). > btrfs filesystem show failed to read /dev/sr0 Label: none uuid: 78b5162b-489e-4de1-a989-a47b91adef50 Total devices 2 FS bytes used 107.64GB devid 2 size 149.05GB used 109.01GB path /dev/sdh1 devid 1 size 156.81GB used 109.03GB path /dev/sdb6 Btrfs v0.20-rc1 >
2013 Oct 04
1
btrfs raid0
How can I verify the read speed of a btrfs raid0 pair in archlinux.? I assume raid0 means striped activity in a paralleled mode at lease similar to raid0 in mdadm. How can I measure the btrfs read speed since it is copy-on-write which is not the norm in mdadm raid0.? Perhaps I cannot use the same approach in btrfs to determine the performance. Secondly, I see a methodology for raid10 using
2012 Oct 25
46
[RFC] New attempt to a better "btrfs fi df"
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi all, this is a new attempt to improve the output of the command "btrfs fi df". The previous attempt received a good reception. However there was no a general consensus about the wording. Moreover I still didn''t understand how btrfs was using the disks. A my first attempt was to develop a new command which shows how the disks
2013 May 10
5
Btrfs balance invalid argument error
Hi list, I am using kernel 3.9.0, btrfs-progs 0.20-rc1-253-g7854c8b. I have a three disk array of level single: # btrfs fi sh Label: none uuid: 2e905f8f-e525-4114-afa6-cce48f77b629 Total devices 3 FS bytes used 3.80TB devid 1 size 2.73TB used 2.25TB path /dev/sdd devid 2 size 2.73TB used 1.55TB path /dev/sdc devid 3 size 2.73TB used 0.00 path /dev/sdb
2012 Jan 17
8
[RFC][PATCH 1/2] Btrfs: try to allocate new chunks with degenerated profile
If there is no free space, the free space allocator will try to get space from the block group with the degenerated profile. For example, if there is no free space in the RAID1 block groups, the allocator will try to allocate space from the DUP block groups. And besides that, the space reservation has the similar behaviour: if there is no enough space in the space cache to reserve, it will reserve
2009 Sep 24
5
OT: What's wrong with RAID5
Hi all, Sorry for the OT. I've got an IBM N3300-A10 NAS. It runs Data Ontap 7.2.5.1. The problem is, from the docs it says that it only supports either RAID-DP or RAID4. What I want to achieve is Max Storage Capacity, so I change it from RAID-DP to RAID4, but with RAID4, the maximum disk in a RAID Group decrease from 14 to 7. In the end, either using RAID-DP or RAID4, the capacity is the same.
2011 Nov 23
2
stripe alignment consideration for btrfs on RAID5
Hiya, is there any recommendation out there to setup a btrfs FS on top of hardware or software raid5 or raid6 wrt stripe/stride alignment? From mkfs.btrfs, it doesn''t look like there''s much that can be adjusted that would help, and what I''m asking might not even make sense for btrfs but I thought I''d just ask. Thanks, Stephane -- To unsubscribe from this
2011 May 04
2
Cannot resize btrfs volume
Hello, I added a new disk into our RAID5 array, it looks like this: md2 : active raid5 sdd4[3] sde4[4] sda4[0] sdc4[2] sdb4[1] 3767274240 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU] # btrfs fi sh Label: none uuid: 5534d2e7-be31-49c7-8ab7-90c5ab8afe18 Total devices 1 FS bytes used 2.24TB devid 3 size 2.63TB used 2.63TB path /dev/md2 # mount ... /dev/md2 on /home type btrfs
2011 Dec 28
3
Btrfs: blocked for more than 120 seconds, made worse by 3.2 rc7
Hello all: I have two machines with btrfs, that give me the "blocked for more than 120 seconds" message. After that I cannot write anything to disk, i am unable to unmount the btrfs filesystem and i can only reboot with sysrq-trigger. It always happens when i write many files with rsync over network. When i used 3.2rc6 it happened randomly on both machines after 50-500gb of writes.
2012 May 07
53
kernel 3.3.4 damages filesystem (?)
Hallo, "never change a running system" ... For some months I run btrfs unter kernel 3.2.5 and 3.2.9, without problems. Yesterday I compiled kernel 3.3.4, and this morning I started the machine with this kernel. There may be some ugly problems. Copying something into the btrfs "directory" worked well for some files, and then I got error messages (I''ve not
2020 Sep 10
3
Btrfs RAID-10 performance
I cannot verify it, but I think that even JBOD is propagated as a virtual device. If you create JBOD from 3 different disks, low level parameters may differ. And probably old firmware is the reason we used RAID-0 two or three years before. Thank you for the ideas. Kind regards Milo Dne 10.09.2020 v 16:15 Scott Q. napsal(a): > Actually there is, filesystems like ZFS/BTRFS prefer to see
2012 Feb 07
2
Understanding Default RAID Behavior
The Wiki does not make it clear as to why adding a secondary device defaults to RAID1 metadata and RAID0 data. I bought two SSDs with the intention of doing a BTRFS RAID0 for my root. What is the difference between forcing RAID0 on metadata and data as opposed to the default behavior? Can anyone clarify that? Thank you for your time, Mario -- To unsubscribe from this list: send the line
2011 Jan 18
6
BUG while writing to USB btrfs filesystem
While untar''ing an image to an sd card via a reader, I got the following bug. The system also has a btrfs root, and a whole swath of processes went into uninterruptable sleep. I was able to poke around via ssh and sysrq, and already had netconsole set up to capture the bug. Root fs is on /dev/sdi1, and /dev/sdj2 is the card reader which was the target of the untar. [29571.448889] sd
2011 Dec 07
7
FS won't mount, open_ctree failed, Assertion !(path->slots[0] == 0) failed
So I''m having a bit of trouble with one of my btrfs filesystems. It isn''t mounting after a power failure. I can''t get restore or btrfsck to run, even on backup supers. I''ve pasted some output below. My btrfs-progs below come from git this morning, running on linux 3.2.0. # mount /dev/md2 /media/test/ mount: wrong fs type, bad option, bad superblock on
2012 Apr 02
2
[PATCH 0/2] Fix btrfs blocksize and bind mkfs.btrfs (RHBZ#807905).
https://bugzilla.redhat.com/show_bug.cgi?id=807905 Currently if you specify the blocksize parameter to mkfs-opts with a btrfs filesystem, then it fails, because mkfs.btrfs interprets the -b option as meaning filesystem size. The first patch fixes this by disallowing blocksize (it cannot be mapped meaningfully into btrfs parameters). The second patch adds the full /sbin/mkfs.btrfs utility to the