Displaying 20 results from an estimated 10000 matches similar to: "how to relocate a disk"
2007 Aug 14
2
restore lost pool after vtoc re-label
hi all,
i''ve been using a SAN LUN as the sole member of a zpool with one additional zfs filesystem. this is a flat SAN fabric, so this LUN was available to other systems on the fabric, and one of them came up with "wrong magic number" for several drives, and, as best i can tell, the vtoc for my zpool LUN was over-written on that host via format labeling to correct the error.
2007 Feb 03
4
Which label a ZFS/ZPOOL device has ? VTOC or EFI ?
Hi All,
ZPOOL / ZFS commands writes EFI label on a device if we create ZPOOL/ZFS fs on it. Is it true ?
I formatted a device with VTOC lable and I created a ZFS file system on it.
Now which label the ZFS device has ? is it old VTOC or EFI ?
After creating the ZFS file system on a VTOC labeled disk, I am seeing the following warning messages.
Feb 3 07:47:00 scoobyb
2006 Nov 03
27
# devices in raidz.
for s10u2, documentation recommends 3 to 9 devices in raidz. what is the
basis for this recommendation? i assume it is performance and not failure
resilience, but i am just guessing... [i know, recommendation was intended
for people who know their raid cold, so it needed no further explanation]
thanks... oz
--
ozan s. yigit | oz at somanetworks.com | 416 977 1414 x 1540
I have a hard time
2006 Mar 10
3
pool space reservation
What is a use case of setting a reservation on the base pool object?
Say I have a pool of 3 100GB drives dynamic striped (pool size of 300GB), and I set the reservation to 200GB. I don''t see any commands that let me ever reduce a pool''s size, so how is the 200GB reservation used?
Related question: is there a plan in the future to allow me to replace the 3 100GB drives with 2
2007 Sep 17
2
zpool create -f not applicable to hot spares
Hello zfs-discuss,
If you do ''zpool create -f test A B C spare D E'' and D or E contains
UFS filesystem then despite of -f zpool command will complain that
there is UFS file system on D.
workaround: create a test pool with -f on D and E, destroy it and
that create first pool with D and E as hotspares.
I''ve tested it on s10u3 + patches - can someone confirm
2008 Jan 23
4
Synchronous scrub?
Say I''m firing off an at(1) or cron(1) job to do scrubs, and say I want to scrub two pools sequentially
because they share one device. The first pool, BTW, is a mirror comprising of a smaller disk and a subset of a larger disk. The other pool is the remainder of the larger disk.
I see no documentation mentioning how to scrub, then wait-until-completed. I''m happy to be pointed
2008 Apr 12
5
ZVOL access permissions?
How can I set up a ZVOL that''s accessible by non-root users, too? The intent is to use sparse ZVOLs as raw disks in virtualization (reducing overhead compared to file-based virtual volumes).
Thanks,
-mg
This message posted from opensolaris.org
2008 Apr 29
4
Finding Pool ID
Folks,
How can I find out zpool id without using zpool import? zpool list
and zpool status does not have option as of Solaris 10U5.. Any back door
to grab this property will be helpful.
Thank you
Ajay
2006 May 16
3
ZFS snv_b39 and S10U2
Hello zfs-discuss,
Just to be sure - if I create ZFS filesystems on snv_39 and then
later I would want just to import that pool on S10U2 - can I safely
assume it will just work (I mean nothing new to on-disk format was
added or changed in last few snv releases which is not going to be
in u2)?
I want to put right now (I have to do it now) some data on ZFS and
later I want to
2009 Mar 27
7
is zpool export/import | faster than rsync or cp
I need to move data from one zpool to another, lock stock and
barrel,
Being from linux background my instinct was to use rsync. But then I
remembered seeing the `export/import options in man zpool.. And I''ve
seen mention of them here too, but didn''t pay attention since I''d
noticed no need yet.
Now I''m wondering if the export/import sub commands might not be
2006 Mar 30
39
Proposal: ZFS Hot Spare support
As mentioned last night, we''ve been reviewing a proposal for hot spare
support in ZFS. Below you can find a current draft of the proposed
interfaces. This has not yet been submitted for ARC review, but
comments are welcome. Note that this does not include any enhanced FMA
diagnosis to determine when a device is "faulted". This will come in a
follow-on project, of which some
2008 Jan 15
4
Moving zfs to an iscsci equallogic LUN
We have a Mirror setup in ZFS that''s 73GB (two internal disks on a sun fire v440). We currently are going to attach this system to an Equallogic Box, and will attach an ISCSCI LUN from the Equallogic box to the v440 of about 200gb. The Equallogic Box is configured as a Hardware Raid 50 (two hot spares for redundancy).
My question is what''s the best approach to moving the ZFS
2007 Apr 22
7
slow sync on zfs
Hello zfs-discuss,
Relatively low traffic to the pool but sync takes too long to complete
and other operations are also not that fast.
Disks are on 3510 array. zil_disable=1.
bash-3.00# ptime sync
real 1:21.569
user 0.001
sys 0.027
During sync zpool iostat and vmstat look like:
f3-1 504G 720G 370 859 995K 10.2M
misc 20.6M 52.0G 0 0
2009 Dec 03
5
L2ARC in clusters
Hi,
When deploying ZFS in cluster environment it would be nice to be able
to have some SSDs as local drives (not on SAN) and when pool switches
over to the other node zfs would pick up the node''s local disk drives as
L2ARC.
To better clarify what I mean lets assume there is a 2-node cluster with
1sx 2540 disk array.
Now lets put 4x SSDs in each node (as internal/local drives). Now
2007 Feb 05
6
snapdir visable recursively throughout a dataset
Is there an existing RFE for, what I''ll wrongly call, "recursively visable snapshots"? That is, .zfs in directories other than the dataset root.
Frankly, I don''t need it available in all directories, although it''d be nice, but I do have a need for making it visiable 1 dir down from the dataset root. The problem is that while ZFS and Zones work smoothly
2006 Jun 12
3
zfs destroy - destroying a snapshot
Hello zfs-discuss,
I''m writing a script to do automatically snapshots and destroy old
one. I think it would be great to add to zfs destroy another option
so only snapshots can be destroyed. Something like:
zfs destroy -s SNAPSHOT
so if something other than snapshot is provided as an argument
zfs destroy wouldn''t actually destroy it.
That way it would
2007 Dec 17
1
HA-NFS AND HA-ZFS
We are currently running sun cluster 3.2 on solaris 10u3. We are using ufs/vxvm 4.1 as our shared file systems. However, I would like to migrate to HA-NFS on ZFS. Since there is no conversion process from UFS to ZFS other than copy, I would like to migrate on my own time. To do this I am planning to add a new zpool HAStoragePlus resource to my existing HA-NFS resource group. This way I can migrate
2008 Jan 25
1
Order of operations w/ checksum errors
zpool status shows a few checksum errors against 1 device in a raidz1 3 disk array and no read or write errors against that device. The pool marked as degraded. Is there a difference if you clear the errors for the pool before you scrub versus scrubing then clearing the errors? I''m not sure if the clearing errors prior to a scrub will replicate out any bad blocks that were identified as
2007 Mar 16
8
ZFS checksum error detection
Hi all.
A quick question about the checksum error detection routines in ZFS.
Surely ZFS can decide about checksum errors in a redundant environment but
what about an non-redundant one? We connected a single RAID5 array to a
v440 as a NFS server and while doing backups and the like we see the
"zpool status -v" checksum error counters increment once in a while.
Nevertheless the
2007 Sep 18
3
ZFS and encryption
Hello zfs-discuss,
I wonder if ZFS will be able to take any advantage of Niagara''s
built-in crypto?
--
Best regards,
Robert Milkowski mailto:rmilkowski at task.gda.pl
http://milek.blogspot.com