similar to: Mirrrors with Uneven Drives!?

Displaying 20 results from an estimated 3000 matches similar to: "Mirrrors with Uneven Drives!?"

2010 Feb 20
6
l2arc current usage (population size)
Hello, How do you tell how much of your l2arc is populated? I''ve been looking for a while now, can''t seem to find it. Must be easy, as this blog entry shows it over time: http://blogs.sun.com/brendan/entry/l2arc_screenshots And follow up, can you tell how much of each data set is in the arc or l2arc? -- This message posted from opensolaris.org
2007 Sep 18
5
ZFS panic in space_map.c line 125
One of our Solaris 10 update 3 servers paniced today with the following error: Sep 18 00:34:53 m2000ef savecore: [ID 570001 auth.error] reboot after panic: assertion failed: ss != NULL, file: ../../common/fs/zfs/space_map.c, line: 125 The server saved a core file, and the resulting backtrace is listed below: $ mdb unix.0 vmcore.0 > $c vpanic() 0xfffffffffb9b49f3() space_map_remove+0x239()
2006 Nov 09
16
Some performance questions with ZFS/NFS/DNLC at snv_48
Hello. We''re currently using a Sun Blade1000 (2x750MHz, 1G ram, 2x160MB/s mpt scsi buses, skge GigE network) as a NFS backend with ZFS for distribution of free software like Debian (cdimage.debian.org, ftp.se.debian.org) and have run into some performance issues. We are running SX snv_48 and have run with a raidz2 with 7x300G for a while now, just added another 7x300G raidz2 today but
2008 Jun 07
4
Mixing RAID levels in a pool
Hi, I had a plan to set up a zfs pool with different raid levels but I ran into an issue based on some testing I''ve done in a VM. I have 3x 750 GB hard drives and 2x 320 GB hard drives available, and I want to set up a RAIDZ for the 750 GB and mirror for the 320 GB and add it all to the same pool. I tested detaching a drive and it seems to seriously mess up the entire pool and I
2009 Apr 08
2
ZFS data loss
Hi, I have lost a ZFS volume and I am hoping to get some help to recover the information ( a couple of months worth of work :( ). I have been using ZFS for more than 6 months on this project. Yesterday I ran a "zvol status" command, the system froze and rebooted. When it came back the discs where not available. See bellow the output of " zpool status", "format"
2008 Jun 09
2
creating ZFS mirror over iSCSI between to DELL MD3000i arrays
Hi, I''ve looked at ZFS for a while now and i''m wondering if it''s possible on a server create a ZFS mirror between two different iSCSI targets (two MD3000i located in two different server rooms). Or is it any setup that you guys recommend for maximal data protection. Thanks, /Thom This message posted from opensolaris.org
2007 Aug 27
1
Nested ZFS sharenfs exports are empty on automount clients
Hello I''ve got nested ZFS filesystems exported via NFS. They are mounted on the clients using automount (from a NIS map). But: only the root exported filesystem shows any contents on the clients. Any sub-directories it has are fine, but any sub-filesystems are empty. ie. NIS map auto.stuff contains "stuff server:/stuff/images" server% zfs get sharenfs stuff/images
2010 Aug 03
2
When is the L2ARC refreshed if on a separate drive?
I''m running a mirrored pair of 2 TB SATA drives as my data storage drives on my home workstation, a Core i7-based machine with 10 GB of RAM. I recently added a sandforce-based 60 GB SSD (OCZ Vertex 2, NOT the pro version) as an L2ARC to the single mirrored pair. I''m running B134, with ZFS pool version 22, with dedup enabled. If I understand correctly, the dedup table should be in
2010 Apr 02
6
L2ARC & Workingset Size
Hi all I ran a workload that reads & writes within 10 files each file is 256M, ie, (10 * 256M = 2.5GB total Dataset Size). I have set the ARC max size to 1 GB on etc/system file In the worse case, let us assume that the whole dataset is hot, meaning my workingset size= 2.5GB My SSD flash size = 8GB and being used for L2ARC No slog is used in the pool My File system record size = 8K ,
2008 Jan 10
2
NCQ
fun example that shows NCQ lowers wait and %w, but doesn''t have much impact on final speed. [scrubbing, devs reordered for clarity] extended device statistics device r/s w/s kr/s kw/s wait actv svc_t %w %b sd2 454.7 0.0 47168.0 0.0 0.0 5.7 12.6 0 74 sd4 440.7 0.0 45825.9 0.0 0.0 5.5 12.4 0 78 sd6 445.7 0.0
2009 Jun 19
8
x4500 resilvering spare taking forever?
I''ve got a Thumper running snv_57 and a large ZFS pool. I recently noticed a drive throwing some read errors, so I did the right thing and zfs replaced it with a spare. Everything went well, but the resilvering process seems to be taking an eternity: # zpool status pool: bigpool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was
2008 Feb 21
3
raidz2 resilience on 3 disks
Hello, 1) If i create a raidz2 pool on some disks, start to use it, then the disks'' controllers change. What will happen to my zpool? Will it be lost or is there some disk tagging which allows zfs to recognise the disks? 2) if i create a raidz2 on 3 HDs, do i have any resilience? If any one of those drives fails, do i loose everything? I''ve got one such pool and
2009 Feb 05
1
nfs sharing of zfs sub filesystems - can it be done?
I''m new to zfs and opensolaris and so am not sure of the correct terminology for this question. I have a test machine running opensolaris 2008.11 (which I have been very impressed with so far). It has 1 disc for boot and 3 as a zpool (called tank, as per the majority of the examples ;)) This machine is a test "SAN" for use by a second test machine running vmware ESX (the free
2010 Dec 09
3
How many files & directories in a ZFS filesystem?
Looking for a little help, please. A contact from Oracle (Sun) suggested I pose the question to this email. We''re using ZFS on Solaris 10 in an application where there are so many directory-subdirectory layers, and a lot of small files (~1-2Kb) that we ran out of inodes (over 30 million!). So, the zfs question is, how can we see how many files & directories have been created in
2009 Mar 09
3
cannot mount ''/export'' directory is not empty
Hello, I am desperate. Today I realized that my OS 108 doesn''t want to boot. I have no idea what I screwed up. I upgraded on 108 last week without any problems. Here is where I''m stuck: Reading ZFS config: done. Mounting ZFS filesystems: (1/17) cannot mount ''/export'': directory is not empty (17/17) $ svcs -x svc:/system/filesystem/local:default (local file
2010 Feb 18
3
improve meta data performance
We have a SunFire X4500 running Solaris 10U5 which does about 5-8k nfs ops of which about 90% are meta data. In hind sight it would have been significantly better to use a mirrored configuration but we opted for 4 x (9+2) raidz2 at the time. We can not take the downtime necessary to change the zpool configuration. We need to improve the meta data performance with little to no money. Does anyone
2009 Feb 17
5
scrub on snv-b107
scrub completed after 1h9m with 0 errors on Tue Feb 17 12:09:31 2009 This is about twice as slow as the same srub on a solaris 10 box with a mirrored zfs root pool. Has scrub become that much slower? And if so, why? -- Dick Hoogendijk -- PGP/GnuPG key: 01D2433D + http://nagual.nl/ | SunOS sxce snv107 ++ + All that''s really worth doing is what we do for others (Lewis Carrol)
2008 Dec 19
4
ZFS boot and data on same disk - is this supported?
I have read the ZFS best practice guide located at http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide However I have questions whether we support using slices for data on the same disk as we use for ZFS boot. What issues does this create if we have a disk failure in a mirrored environment? Does anyone have examples of customers doing this in production environments. I
2009 Jan 05
3
ZFS import on pool with same name?
I have an OpenSolaris snv_101 box with ZFS on it. (Sun Ultra 20 M2) zpool name is rpool. The I have a 2nd hard drive in the box that I am trying to recover the ZFS data from (long story but that HD became unbootable after installing IPS on the machine) Both drives have a pool named "rpool", so I can''t import the rpool from the 2nd drive. root at hyperion:~# zpool status
2008 Oct 11
5
questions about replacing a raidz2 vdev disk with a larger one
I''d like to replace/upgrade two 500GB disks in RaidZ2 vdev with 1TB disks, but I have some preliminary questions/concerns before trying ''zfs replace dpool ?'' Will ZFS permit this replacement? Will ZFS use the extra space in a heterogeneous RaidZ2 vdev, or is the size limited by the smallest disk in the vdev? Thanks in advance, Vizzini The system is currently running