similar to: Can ZFS simply concatenate LUNs (eg no RAID0)?

Displaying 20 results from an estimated 6000 matches similar to: "Can ZFS simply concatenate LUNs (eg no RAID0)?"

2010 Mar 26
23
RAID10
Hi All, I am looking at ZFS and I get that they call it RAIDZ which is similar to RAID 5, but what about RAID 10? Isn''t a RAID 10 setup better for data protection? So if I have 8 x 1.5tb drives, wouldn''t I: - mirror drive 1 and 5 - mirror drive 2 and 6 - mirror drive 3 and 7 - mirror drive 4 and 8 Then stripe 1,2,3,4 Then stripe 5,6,7,8 How does one do this with ZFS?
2020 Nov 05
1
BIOS RAID0 and differences between disks
> On Nov 4, 2020, at 9:21 PM, John Pierce <jhn.pierce at gmail.com> wrote: > > is it RAID 0 (striped) or raid1 (mirrored) ?? > > if you wrote on half of a raid0 stripe set, you basically trashed it. > blocks are striped across both drives, so like 16k on the first disk, then > 16k on the 2nd then 16k back on the first, repeat (replace 16k with > whatever your raid
2011 Jul 15
22
Zil on multiple usb keys
This might be a stupid question, but here goes... Would adding, say, 4 4 or 8gb usb keys as a zil make enough of a difference for writes on an iscsi shared vol? I am finding reads are not too bad (40is mb/s over gige on 2 500gb drives stripped) but writes top out at about 10 and drop a lot lower... If I where to add a couple usb keys for zil, would it make a difference? Thanks. Sent from a
2013 Oct 04
1
btrfs raid0
How can I verify the read speed of a btrfs raid0 pair in archlinux.? I assume raid0 means striped activity in a paralleled mode at lease similar to raid0 in mdadm. How can I measure the btrfs read speed since it is copy-on-write which is not the norm in mdadm raid0.? Perhaps I cannot use the same approach in btrfs to determine the performance. Secondly, I see a methodology for raid10 using
2020 Nov 05
3
BIOS RAID0 and differences between disks
My computer running CentOS 7 is configured to use BIOS RAID0 and has two identical SSDs which are also encrypted. I had a crash the other day and due to a bug in the operating system update, I am unable to boot the system in RAID mode since dracut does not recognize the disks in grub. After modifying the grub command line I am able to boot the system from one of the harddisks after entering the
2020 Nov 12
1
BIOS RAID0 and differences between disks
On 11/04/2020 10:21 PM, John Pierce wrote: > is it RAID 0 (striped) or raid1 (mirrored) ?? > > if you wrote on half of a raid0 stripe set, you basically trashed it. > blocks are striped across both drives, so like 16k on the first disk, then > 16k on the 2nd then 16k back on the first, repeat (replace 16k with > whatever your raid stripe size is). > > if its a raid 1
2017 Oct 17
1
lvconvert(split) - raid10 => raid0
hi guys, gals do you know if conversion from lvm's raid10 to raid0 is possible? I'm fiddling with --splitmirrors but it gets me nowhere. On "takeover" subject man pages says: "..between striped/raid0 and raid10."" but no details, nowhere I could find documentation, nor a howto. many thanks, L.
2007 Dec 23
11
RAIDZ(2) expansion?
I skimmed the archives and found a thread from July earlier this year about RAIDZ expansion. Not adding more RAIDZ stripes to a pool, but adding more drives to the stripe itself. I''m wondering if an RFE has been submitted for this and if any progress has been made, or is expected? I find myself out of space on my current RAID5 setup and would love to flip over to a ZFS raidz2 solution
2006 Nov 28
7
Convert Zpool RAID Types
Hello, Is it possible to non-destructively change RAID types in zpool while the data remains on-line? -J
2015 Feb 16
4
Centos 7.0 and mismatched swap file
On Mon, Feb 16, 2015 at 6:47 AM, Eliezer Croitoru <eliezer at ngtech.co.il> wrote: > I am unsure I understand what you wrote. > "XFS will create multiple AG's across all of those > devices," > Are you comparing md linear/concat to md raid0? and that the upper level XFS > will run on top them? Yes to the first question, I'm not understanding the second
2007 Apr 18
2
zfs block allocation strategy
Hi, quoting from zfs docs "The SPA allocates blocks in a round-robin fashion from the top-level vdevs. A storage pool with multiple top-level vdevs allows the SPA to use dynamic striping to increase disk bandwidth. Since a new block may be allocated from any of the top-level vdevs, the SPA implements dynamic striping by spreading out writes across all available top-level vdevs" Now,
2007 Sep 11
7
compression=on and zpool attach
I''ve got 12Gb or so of db+web in a zone on a ZFS filesystem on a mirrored zpool. Noticed during some performance testing today that its i/o bound but using hardly any CPU, so I thought turning on compression would be a quick win. I know I''ll have to copy files for existing data to be compressed, so I was going to make a new filesystem, enable compression and rysnc everything in,
2007 Sep 27
6
Best option for my home file server?
I was recently evaluating much the same question but with out only a single pool and sizing my disks equally. I only need about 500GB of usable space and so I was considering the value of 4x 250GB SATA Drives versus 5x 160GB SATA drives. I had intended to use an AMS 5 disk in 3 5.25" bay hot-swap backplane. http://www.american-media.com/product/backplane/sata300/sata300.html I priced
2009 Nov 27
5
unexpected raid1 behavior?
Hi, I''m starting to play with btrfs on my new computer. I''m running Gentoo and have compiled the 2.6.31 kernel, enabling btrfs. Now I have 2 partitions (on 2 different sata disks) that are free for me to play with, each about 375 gb in size. I wanted to create a "raid1" volume using these two partitions, so I did: # mkfs.btrfs -d raid1 /dev/sda5 /dev/sdb5 # mount
2013 Jan 30
8
RAID 0 across SSD and HDD
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 I''ve been unable to find anything definitive about what happens if I use RAID0 to join an SSD and HDD together with respect to performance (latency, throughput). The future is obvious (hot data tracking, using most appropriate device for the data, data migration). In my specific case I have a 250GB SSD and a 500GB HDD, and about 250GB of
2011 Mar 01
5
btrfs wishlist
Hi all Having managed ZFS for about two years, I want to post a wishlist. INCLUDED IN ZFS - Mirror existing single-drive filesystem, as in ''zfs attach'' - RAIDz-stuff - single and hopefully multiple-parity RAID configuration with block-level checksumming - Background scrub/fsck - Pool-like management with multiple RAIDs/mirrors (VDEVs) - Autogrow as in ZFS autoexpand NOT
2012 Jan 17
8
[RFC][PATCH 1/2] Btrfs: try to allocate new chunks with degenerated profile
If there is no free space, the free space allocator will try to get space from the block group with the degenerated profile. For example, if there is no free space in the RAID1 block groups, the allocator will try to allocate space from the DUP block groups. And besides that, the space reservation has the similar behaviour: if there is no enough space in the space cache to reserve, it will reserve
2008 Jan 31
16
Hardware RAID vs. ZFS RAID
Hello, I have a Dell 2950 with a Perc 5/i, two 300GB 15K SAS drives in a RAID0 array. I am considering going to ZFS and I would like to get some feedback about which situation would yield the highest performance: using the Perc 5/i to provide a hardware RAID0 that is presented as a single volume to OpenSolaris, or using the drives separately and creating the RAID0 with OpenSolaris and ZFS? Or
2015 Feb 16
2
Centos 7.0 and mismatched swap file
On Mon, Feb 16, 2015 at 12:07 AM, Michael Schumacher <michael.schumacher at pamas.de> wrote: > Btw., are you sure you want to use XFS for a mail server? I made some > tests about a year ago and found that EXT4 is by the factor 10 faster > compared to XFS. The tests I performed were using the "maildir" style > postfix installation that results in many thousands files in
2007 Apr 02
4
Convert raidz
Hi Is it possible to convert live 3 disks zpool from raidz to raidz2 And is it possible to add 1 new disk to raidz configuration without backups and recreating zpool from cratch. Thanks This message posted from opensolaris.org