similar to: Inconcistancies with scrub and zdb

Displaying 20 results from an estimated 2000 matches similar to: "Inconcistancies with scrub and zdb"

2009 Jun 01
7
Does zpool clear delete corrupted files
Hi list, First off: # cat /etc/release Solaris 10 6/06 s10x_u2wos_09a X86 Copyright 2006 Sun Microsystems, Inc. All Rights Reserved. Use is subject to license terms. Assembled 09 June 2006 Here''s an (almost) disaster scenario that came to life over the past week. We have a very large zpool
2007 Dec 15
4
Is round-robin I/O correct for ZFS?
I''m testing an Iscsi multipath configuration on a T2000 with two disk devices provided by a Netapp filer. Both the T2000 and the Netapp have two ethernet interfaces for Iscsi, going to separate switches on separate private networks. The scsi_vhci devices look like this in `format'': 1. c4t60A98000433469764E4A413571444B63d0 <NETAPP-LUN-0.2-50.00GB>
2008 Jan 30
18
ZIL controls in Solaris 10 U4?
Is it true that Solaris 10 u4 does not have any of the nice ZIL controls that exist in the various recent Open Solaris flavors? I would like to move my ZIL to solid state storage, but I fear I can''t do it until I have another update. Heck, I would be happy to just be able to turn the ZIL off to see how my NFS on ZFS performance is effected before spending the $''s. Anyone
2008 Nov 12
6
Inexpensive ZFS home server
For anyone looking for a cheap home ZFS server... Dell is having a sale on their PowerEdge SC440 for $199 (regular $598) through 11/12/2008. http://www.dell.com/content/products/productdetails.aspx/pedge_sc440?c=us&cs=04&l=en&s=bsd Its got Dual Core Intel? Pentium?E2180, 2.0GHz, 1MB Cache, 800MHz FSB and you can upgrade the memory (ECC too) to 2gb for 19$ bucks. @$199, I just
2009 Aug 02
2
zdb assertion failure/zpool recovery
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Hi, I have a corrupt pool, which lives on a .vdi file of a VirtualBox. IIRC the corruption (i.e. pool being not importable) was caused when I killed virtual box, because it was hung. This pool consists of a single vdev and I would really like to get some files out of that thing. So I tried running zdb, but this fails with an assertion failure:
2007 Apr 14
1
Move data from the zpool (root) to a zfs file system
Hi List, As a ZFS newbie, I foolishly copied my data set to the root zpool file system (a large iSCSI SAN array). Thus: # zpool create -f iscsi c4t19d0 c4t20d0 c4t21d0 c4t22d0 c4t23d0 c4t24d0 # zpool list NAME SIZE USED AVAIL CAP HEALTH ALTROOT iscsi 9.53T 64.5K 5.34T 0% ONLINE - # zfs set mountpoint=/mydisks/iscsi iscsi Then copied
2010 Aug 30
5
pool died during scrub
I have a bunch of sol10U8 boxes with ZFS pools, most all raidz2 8-disk stripe. They''re all supermicro-based with retail LSI cards. I''ve noticed a tendency for things to go a little bonkers during the weekly scrub (they all scrub over the weekend), and that''s when I''ll lose a disk here and there. OK, fine, that''s sort of the point, and they''re
2007 Feb 11
0
unable to mount legacy vol - panic in zfs:space_map_remove - zdb crashes
I have a 100gb SAN lun in a pool, been running ok for about 6 months. panicked the system this morning. system was running S10U2. In the course of troubleshooting I''ve installed the latest recommended bundle including kjp 118833-36 and zfs patch 124204-03 created as: zpool create zfspool01 /dev/dsk/emcpower0c zfs create zfspool01/nb60openv zfs set mountpoint=legacy zfspool01/nb60openv
2007 Sep 13
26
hardware sizing for a zfs-based system?
Hi all, I''m putting together a OpenSolaris ZFS-based system and need help picking hardware. I''m thinking about using this 26-disk case: [FYI: 2-disk RAID1 for the OS & 4*(4+2) RAIDZ2 for SAN] http://rackmountpro.com/productpage.php?prodid=2418 Regarding the mobo, cpus, and memory - I searched goggle and the ZFS site and all I came up with so far is that, for a
2007 Sep 13
26
hardware sizing for a zfs-based system?
Hi all, I''m putting together a OpenSolaris ZFS-based system and need help picking hardware. I''m thinking about using this 26-disk case: [FYI: 2-disk RAID1 for the OS & 4*(4+2) RAIDZ2 for SAN] http://rackmountpro.com/productpage.php?prodid=2418 Regarding the mobo, cpus, and memory - I searched goggle and the ZFS site and all I came up with so far is that, for a
2012 Jan 17
0
ZDB returning strange values
Hello all, I have a question about what output "ZDB -dddddd" should produce in L0 DVA fields. I expected there to be one or more same-sized references to data blocks stored in top-level vdevs (one vdev #0 in my 6-disk raidz2 pool), as confirmed by the source: http://src.illumos.org/source/xref/illumos-gate/usr/src/cmd/zdb/zdb.c#sprintf_blkptr_compact And I do see that for some of my
2009 Dec 15
7
ZFS Dedupe reporting incorrect savings
Hi, Created a zpool with 64k recordsize and enabled dedupe on it. zpool create -O recordsize=64k TestPool device1 zfs set dedup=on TestPool I copied files onto this pool over nfs from a windows client. Here is the output of zpool list Prompt:~# zpool list NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT TestPool 696G 19.1G 677G 2% 1.13x ONLINE - When I ran a
2012 Jan 31
0
(gang?)block layout question, and how to decipher ZDB output?
Hello, all I''m "playing" with ZDB again on another test system, the rpool being uncompressed with 512-byte sectors. Here''s some output that puzzles me (questions follow): # zdb -dddddddd -bbbbbb rpool/ROOT/nightly-2012-01-31 260050 ... 1e80000 L0 DVA[0]=<0:200972e00:20200> DVA[1]=<0:391820a00:200> [L0 ZFS plain file] fletcher4 uncompressed
2008 Aug 05
0
mdb & zdb should print info about crypt in blkptr
Author: Darren Moffat <Darren.Moffat at Sun.COM> Repository: /hg/zfs-crypto/gate Latest revision: 7a6ad1928ffa250a595fe19b5eef1923cf2a4c67 Total changesets: 1 Log message: mdb & zdb should print info about crypt in blkptr Files: update: usr/src/cmd/mdb/common/modules/zfs/zfs.c update: usr/src/cmd/zdb/zdb.c
2006 Oct 31
0
6366222 zdb(1M) needs to use largefile primitives when reading label
Author: eschrock Repository: /hg/zfs-crypto/gate Revision: e5f70a6fc5010aa205f244a25a9cdb950e0dae89 Log message: 6366222 zdb(1M) needs to use largefile primitives when reading label 6366267 zfs broke non-BUILD64 compilation of libsec Files: update: usr/src/cmd/zdb/zdb.c update: usr/src/lib/libsec/Makefile
2007 Oct 02
53
Direct I/O ability with zfs?
We are using MySQL, and love the idea of using zfs for this. We are used to using Direct I/O to bypass file system caching (let the DB do this). Does this exist for zfs? This message posted from opensolaris.org
2010 Sep 10
3
zpool upgrade and zfs upgrade behavior on b145
Not sure what the best list to send this to is right now, so I have selected a few, apologies in advance. A couple questions. First I have a physical host (call him bob) that was just installed with b134 a few days ago. I upgraded to b145 using the instructions on the Illumos wiki yesterday. The pool has been upgraded (27) and the zfs file systems have been upgraded (5). chris at bob:~# zpool
2010 Sep 10
3
zpool upgrade and zfs upgrade behavior on b145
Not sure what the best list to send this to is right now, so I have selected a few, apologies in advance. A couple questions. First I have a physical host (call him bob) that was just installed with b134 a few days ago. I upgraded to b145 using the instructions on the Illumos wiki yesterday. The pool has been upgraded (27) and the zfs file systems have been upgraded (5). chris at bob:~# zpool
2007 Sep 13
2
zpool versioning
Hi, I was wondering if anyone would know if this is just an accounting-type error with the recorded "version=" stored on disk, or if there are/could-be any deeper issues with an "upgraded" zpool? I created a pool under a Sol10_x86_u3 install (11/06?), and zdb correctly reported the pool as a "version=3" pool. I reinstalled the OS with a u4 (08/07), ran zpool
2009 Oct 31
1
Kernel panic on zfs import
Hi, I''ve got an OpenSolaris 2009.06 box that will reliably panic whenever I try to import one of my pools. What''s the best practice for recovering (before I resort to nuking the pool and restoring from backup)? There are two pools on the system: rpool and tank. The rpool seems to be fine, since I can boot from a 2009.06 CD and ''zpool import -f rpool''; I can