similar to: Understanding Default RAID Behavior

Displaying 20 results from an estimated 20000 matches similar to: "Understanding Default RAID Behavior"

2013 Apr 03
2
[bug] btrfs fi df doesn't show raid type after balance
Did something break.. ? we are not reporting raid type after balance. ----------- # btrfs fi df /btrfs Data, RAID0: total=2.00GB, used=2.03MB Data: total=8.00MB, used=0.00 System, RAID0: total=16.00MB, used=4.00KB System: total=4.00MB, used=0.00 Metadata, RAID0: total=2.00GB, used=216.00KB Metadata: total=8.00MB, used=4.00KB # btrfs bal /btrfs Done, had to relocate 5 out of 5 chunks # btrfs fi
2013 Jan 30
8
RAID 0 across SSD and HDD
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 I''ve been unable to find anything definitive about what happens if I use RAID0 to join an SSD and HDD together with respect to performance (latency, throughput). The future is obvious (hot data tracking, using most appropriate device for the data, data migration). In my specific case I have a 250GB SSD and a 500GB HDD, and about 250GB of
2009 Nov 19
10
Unable to mount loopback devices in RAID mode
Hi! I recently tried to mount a filesystem in RAID1 mode using loopback devices. I followed the instructions at [1]. Here''s exactly what I''ve done: $ dd if=/dev/zero of=raid1_0.img bs=1M count=500 $ dd if=/dev/zero of=raid1_1.img bs=1M count=500 $ mkfs.btrfs -m raid1 -d raid1 raid1_0.img raid1_1.img $ losetup /dev/loop0 raid1_0.img $ losetup /dev/loop1 raid1_1.img $ mount -t
2011 Aug 08
7
“bio too big” regression and silent data corruption in 3.0
tl;dr version: 3.0 produces “bio too big” dmesg entries and silently corrupts data in “meta-raid1/data-single” configurations on disks with different max_hw_sectors, where 2.6.38 worked fine. tl;dr side-issue: on-line removal of partitions holding “single” data attempts to create raid0 (rather than single) block groups. If it can''t get enough room for raid0 over all remaining disks, it
2012 Jan 17
8
[RFC][PATCH 1/2] Btrfs: try to allocate new chunks with degenerated profile
If there is no free space, the free space allocator will try to get space from the block group with the degenerated profile. For example, if there is no free space in the RAID1 block groups, the allocator will try to allocate space from the DUP block groups. And besides that, the space reservation has the similar behaviour: if there is no enough space in the space cache to reserve, it will reserve
2011 Apr 09
16
wrong values in "df" and "btrfs filesystem df"
Hallo, linux-btrfs, First I create an array of 2 disks with mkfs.btrfs -d raid0 -m raid1 /dev/sdb1 /dev/sdd1 and mount it at /srv/MM. Then I fill it with about 1,6 TByte. And then I add /dev/sde1 via btrfs device add /dev/sde1 /srv/MM btrfs filesystem balance /srv/MM (it run about 20 hours) Then I work on it, copy some new files, delete some old files - all works well. Only df
2011 Jan 07
9
Various Questions
On Fri 07 January 2011 08:14:17 Hubert Kario wrote: > I''d suggest at least > mkfs.btrfs -m raid1 -d raid0 /dev/sdc /dev/sdd > if you really want raid0 I don''t fully understand -m or -d. Why would this make a truer raid0 that with no options? Is it necessary to use fdisk on new drives in creating a BTRFS multi-drive array? Or is this all that''s needed: #
2012 May 17
6
SSD format/mount parameters questions
For using SSDs: Are there any format/mount parameters that should be set for using btrfs on SSDs (other than the "ssd" mount option)? General questions: How long is the ''delay'' for the delayed alloc? Are file allocations aligned to 4kiB boundaries, or larger? What byte value is used to pad unused space? (Aside: For some, the erased state reads all 0x00, and for
2009 Nov 27
5
unexpected raid1 behavior?
Hi, I''m starting to play with btrfs on my new computer. I''m running Gentoo and have compiled the 2.6.31 kernel, enabling btrfs. Now I have 2 partitions (on 2 different sata disks) that are free for me to play with, each about 375 gb in size. I wanted to create a "raid1" volume using these two partitions, so I did: # mkfs.btrfs -d raid1 /dev/sda5 /dev/sdb5 # mount
2009 Feb 13
3
Bonnie++ run with RAID-1 on a single SSD (2.6.29-rc4-224-g4b6136c)
Hi folks, For people who might be interested, here is how btrfs performs with two partitions on a single SSD drive in a RAID-1 mirror. This is on a Dell E4200 with Core 2 Duo U9300 (1.2GHz), 2GB RAM and a Samsung SSD (128GB Thin uSATA SSD). Version 1.03c ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block--
2020 Nov 05
3
BIOS RAID0 and differences between disks
My computer running CentOS 7 is configured to use BIOS RAID0 and has two identical SSDs which are also encrypted. I had a crash the other day and due to a bug in the operating system update, I am unable to boot the system in RAID mode since dracut does not recognize the disks in grub. After modifying the grub command line I am able to boot the system from one of the harddisks after entering the
2013 Aug 19
1
LVM RAID0 and SSD discards/TRIM
I'm trying to work out the kinks of a proprietary, old, and clunky application that runs on CentOS. One of its main problems is that it writes image sequences extremely non-linearly and in several passes, using many CPUs, so the sequences get very fragmented. The obvious solution to this seems to be to use SSDs for its output, and some scripts that will pick up and copy our the sequences
2012 Mar 25
3
attempt to access beyond end of device and livelock
Hi Dongyang, Yan, When testing BTRFS with RAID 0 metadata on linux-3.3, we see discard ranges exceeding the end of the block device [1], potentially causing dataloss; when this occurs, filesystem writeback becomes catatonic due to continual resubmission. The reproducer is quite simple [2]. Hope this proves useful... Thanks, Daniel --- [1] attempt to access beyond end of device ram0: rw=129,
2013 Feb 20
7
Hybrid Storage proposal
Here is a short proposal for the hybrid storage cache idea with introduction/motivation and a bird''s eye view of an approach to implement a hybrid storage cache for btrfs. Please note that there is currently no available patches. We would like to get as much input before as possible before we start designing and implementing a solution. 1. Introduction The emerge of Solid State
2013 Aug 16
3
4 vol raid5 segfault on device delete
I have a 4 device volume with raid5 - trying to remove one of the devices (plenty of free space) and I get an almost immediate segfault. Scrub shows no errors, repair show space cache invalid but nothing else (I remounted with clear cache to be safe). Lots of corrupt on bdev (for 3 out of 4 drives), but I have no file access issues that I know of. Thanks! Output below:
2013 May 30
9
oops at mount
hi All, I''m new on the list. System: Distributor ID: Ubuntu Description: Ubuntu 13.04 Release: 13.04 Codename: raring Linux ctu 3.8.0-19-generic #30-Ubuntu SMP Wed May 1 16:35:23 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux The symptom is the same with Saucy 3.9 kernel. ii btrfs-tools 0.20~git20130524~650e656-0daily13~raring1 amd64 Checksumming Copy on
2011 Jan 18
6
BUG while writing to USB btrfs filesystem
While untar''ing an image to an sd card via a reader, I got the following bug. The system also has a btrfs root, and a whole swath of processes went into uninterruptable sleep. I was able to poke around via ssh and sysrq, and already had netconsole set up to capture the bug. Root fs is on /dev/sdi1, and /dev/sdj2 is the card reader which was the target of the untar. [29571.448889] sd
2013 Oct 04
1
btrfs raid0
How can I verify the read speed of a btrfs raid0 pair in archlinux.? I assume raid0 means striped activity in a paralleled mode at lease similar to raid0 in mdadm. How can I measure the btrfs read speed since it is copy-on-write which is not the norm in mdadm raid0.? Perhaps I cannot use the same approach in btrfs to determine the performance. Secondly, I see a methodology for raid10 using
2012 May 07
53
kernel 3.3.4 damages filesystem (?)
Hallo, "never change a running system" ... For some months I run btrfs unter kernel 3.2.5 and 3.2.9, without problems. Yesterday I compiled kernel 3.3.4, and this morning I started the machine with this kernel. There may be some ugly problems. Copying something into the btrfs "directory" worked well for some files, and then I got error messages (I''ve not
2020 Nov 12
1
BIOS RAID0 and differences between disks
On 11/04/2020 10:21 PM, John Pierce wrote: > is it RAID 0 (striped) or raid1 (mirrored) ?? > > if you wrote on half of a raid0 stripe set, you basically trashed it. > blocks are striped across both drives, so like 16k on the first disk, then > 16k on the 2nd then 16k back on the first, repeat (replace 16k with > whatever your raid stripe size is). > > if its a raid 1