Displaying 20 results from an estimated 2000 matches similar to: "Hardware RAID vs. ZFS RAID"
2010 Mar 24
21
ZFS on a 11TB HW RAID-5 controller
Hello all,
I am a complete newbie to OpenSolaris, and must to setup a ZFS NAS. I do have linux experience, but have never used ZFS. I have tried to install OpenSolaris Developer 134 on a 11TB HW RAID-5 virtual disk, but after the installation I can only use one 2TB disk, and I cannot partition the rest. I realize that maximum partition size is 2TB, but I guess the rest must be usable. For
2010 Feb 08
17
ZFS ZIL + L2ARC SSD Setup
I have some questions about the choice of SSDs to use for ZIL and L2ARC.
I''m trying to build an OpenSolaris iSCSI SAN out of a whitebox system,
which is intended to be used as a backup SAN during storage migration,
so it''s built on a tight budget.
The system currently has 4GB RAM, 3GHz Core2-Quad and 8x 500GB WD REII
SATA HDDs attached to an Areca 8port ARC-1220 controller
2011 May 13
27
Extremely slow zpool scrub performance
Running a zpool scrub on our production pool is showing a scrub rate
of about 400K/s. (When this pool was first set up we saw rates in the
MB/s range during a scrub).
Both zpool iostat and an iostat -Xn show lots of idle disk times, no
above average service times, no abnormally high busy percentages.
Load on the box is .59.
8 x 3GHz, 32GB ram, 96 spindles arranged into raidz zdevs on OI 147.
2006 Sep 28
13
jbod questions
Folks,
We are in the process of purchasing new san/s that our mail server
runs on (JES3). We have moved our mailstores to zfs and continue to
have checksum errors -- they are corrected but this improves on the
ufs inode errors that require system shutdown and fsck.
So, I am recommending that we buy small jbods, do raidz2 and let zfs
handle the raiding of these boxes. As we need more
2010 Mar 03
6
Question about multiple RAIDZ vdevs using slices on the same disk
Hi all :)
I''ve been wanting to make the switch from XFS over RAID5 to ZFS/RAIDZ2 for some time now, ever since I read about ZFS the first time. Absolutely amazing beast!
I''ve built my own little hobby server at home and have a boatload of disks in different sizes that I''ve been using together to build a RAID5 array on Linux using mdadm in two layers; first layer is
2011 Feb 14
3
ZFS and Virtual Disks
Hi I wanted to get some expert advice on this. I have an ordinary hardware
SAN from Promise Tech that presents the LUNs via iSCSI. I would like to use
that if possible with my VMware environment where I run several Solaris /
OpenSolaris virtual machines. My question is regarding the virtual disks.
1. Should I create individual iSCSI LUNs and present those to the VMware
ESXi host as iSCSI storage,
2006 Oct 16
11
Configuring a 3510 for ZFS
Hi folks,
Myself and a colleague are currently involved in a prototyping exercise
to evaluate ZFS against our current filesystem. We are looking at the
best way to arrange the disks in a 3510 storage array.
We have been testing with the 12 disks on the 3510 exported as "nraid"
logical devices. We then configured a single ZFS pool on top of this,
using two raid-z arrays. We are getting
2012 Jun 12
2
lost ZFS pool
hail,
I write just to make sure its dead. I've lost the first disk on a ZFS pool (jbod). Now I can't
mount it with only the second disk. The first disk clicks to death :(
[root@optimus ~]# zpool status
pool: pool
state: UNAVAIL
status: One or more devices could not be opened. There are insufficient
replicas for the pool to continue functioning.
action: Attach the missing
2008 Sep 03
8
SAS or SATA HBA with write cache
Anyone know of a SATA and/or SAS HBA with battery backed write cache?
Seems like using a full-blown RAID controller and exporting each individual drive back to ZFS as a single LUN is a waste of power and $$$. Looking for any thoughts or ideas.
Thanks.
-Matt
--
This message posted from opensolaris.org
2007 May 12
3
zfs and jbod-storage
Hi.
I''m managing a HDS storage system which is slightly larger than 100 TB
and we have used approx. 3/4. We use vxfs. The storage system is
attached to a solaris 9 on sparc via a fiberswitch. The storage is
shared via nfs to our webservers.
If I was to replace vxfs with zfs I could utilize raidz(2) instead of
the built-in hardware raid-controller.
Are there any jbod-only storage
2006 Dec 08
22
ZFS Usage in Warehousing (lengthy intro)
Dear all,
we''re currently looking forward to restructure our hardware environment for
our datawarehousing product/suite/solution/whatever.
We''re currently running the database side on various SF V440''s attached via
dual FC to our SAN backend (EMC DMX3) with UFS. The storage system is
(obviously in a SAN) shared between many systems. Performance is mediocre
in terms
2013 Oct 24
4
ZFS on Linux in production?
We are a CentOS shop, and have the lucky, fortunate problem of having
ever-increasing amounts of data to manage. EXT3/4 becomes tough to
manage when you start climbing, especially when you have to upgrade, so
we're contemplating switching to ZFS.
As of last spring, it appears that ZFS On Linux http://zfsonlinux.org/
calls itself production ready despite a version number of 0.6.2, and
2006 Mar 30
39
Proposal: ZFS Hot Spare support
As mentioned last night, we''ve been reviewing a proposal for hot spare
support in ZFS. Below you can find a current draft of the proposed
interfaces. This has not yet been submitted for ARC review, but
comments are welcome. Note that this does not include any enhanced FMA
diagnosis to determine when a device is "faulted". This will come in a
follow-on project, of which some
2006 May 19
11
tracking error to file
In my testing, I''ve found the following error:
zpool status -v
pool: local
state: ONLINE
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
see: http://www.sun.com/msg/ZFS-8000-8A
scrub: none requested
2007 Sep 26
9
Rule of Thumb for zfs server sizing with (192) 500 GB SATA disks?
I''m trying to get maybe 200 MB/sec over NFS for large movie files (need large capacity to hold all of them). Are there any rules of thumb on how much RAM is needed to handle this (probably RAIDZ for all the disks) with zfs, and how large a server should be used? The throughput required is not so large, so I am thinking an X4100 M2 or X4150 should be plenty.
This message posted from
2009 Apr 27
23
Raidz vdev size... again.
Hi,
i''m new to the list so please bare with me. This isn''t an OpenSolaris
related problem but i hope it''s still the right list to post to.
I''m on the way to move a backup server to using zfs based storage, but i
don''t want to spend too much drives to parity (the 16 drives are attached
to a 3ware raid controller so i could also just use raid6 there).
I
2009 Mar 01
8
invalid vdev configuration after power failure
What does it mean for a vdev to have an invalid configuration and how
can it be fixed or reset? As you can see, the following pool can no
longer be imported: (Note that the "last accessed by another system"
warning is because I moved these drives to my test workstation.)
~$ zpool import -f pool0
cannot import ''pool0'': invalid vdev configuration
~$ zpool import
pool:
2010 Jul 20
16
zfs raidz1 and traditional raid 5 perfomrance comparision
Hi,
for zfs raidz1, I know for random io, iops of a raidz1 vdev eqaul to one physical disk iops, since raidz1 is like raid5 , so is raid5 has same performance like raidz1? ie. random iops equal to one physical disk''s ipos.
Regards
Victor
--
This message posted from opensolaris.org
2012 Nov 22
19
ZFS Appliance as a general-purpose server question
A customer is looking to replace or augment their Sun Thumper
with a ZFS appliance like 7320. However, the Thumper was used
not only as a protocol storage server (home dirs, files, backups
over NFS/CIFS/Rsync), but also as a general-purpose server with
unpredictably-big-data programs running directly on it (such as
corporate databases, Alfresco for intellectual document storage,
etc.) in order to
2012 Aug 07
4
LSI 9240-4i 4K alignment
Hi all,
We have a server with a LSI 9240-4i controller configured in JBOD with
4 SATA disks. Running FreeBSD 9.1-Beta1:
Relevant dmesg:
FreeBSD 9.1-BETA1 #0: Thu Jul 12 09:38:51 UTC 2012
root@farrell.cse.buffalo.edu:/usr/obj/usr/src/sys/GENERIC amd64
CPU: Intel(R) Xeon(R) CPU E31230 @ 3.20GHz (3200.09-MHz K8-class CPU)
Origin = "GenuineIntel" Id = 0x206a7 Family = 6 Model =