similar to: ZFS Dataset lost structure

Displaying 20 results from an estimated 100 matches similar to: "ZFS Dataset lost structure"

2012 Dec 20
3
Pool performance when nearly full
Hi I know some of this has been discussed in the past but I can''t quite find the exact information I''m seeking (and I''d check the ZFS wikis but the websites are down at the moment). Firstly, which is correct, free space shown by "zfs list" or by "zpool iostat" ? zfs list: used 50.3 TB, free 13.7 TB, total = 64 TB, free = 21.4% zpool iostat: used
2012 Jan 17
0
ZDB returning strange values
Hello all, I have a question about what output "ZDB -dddddd" should produce in L0 DVA fields. I expected there to be one or more same-sized references to data blocks stored in top-level vdevs (one vdev #0 in my 6-disk raidz2 pool), as confirmed by the source: http://src.illumos.org/source/xref/illumos-gate/usr/src/cmd/zdb/zdb.c#sprintf_blkptr_compact And I do see that for some of my
2012 Jan 31
0
(gang?)block layout question, and how to decipher ZDB output?
Hello, all I''m "playing" with ZDB again on another test system, the rpool being uncompressed with 512-byte sectors. Here''s some output that puzzles me (questions follow): # zdb -dddddddd -bbbbbb rpool/ROOT/nightly-2012-01-31 260050 ... 1e80000 L0 DVA[0]=<0:200972e00:20200> DVA[1]=<0:391820a00:200> [L0 ZFS plain file] fletcher4 uncompressed
2011 Nov 08
1
Single-disk rpool with inconsistent checksums, import fails
Hello all, I have an oi_148a PC with a single root disk, and since recently it fails to boot - hangs after the copyright message whenever I use any of my GRUB menu options. Booting with an oi_148a LiveUSB I had around since installation, I ran some zdb traversals over the rpool and zpool import attempts. The imports fail by running the kernel out of RAM (as recently discussed in the list with
2011 Apr 28
4
Finding where dedup''d files are
Is there an easy way to find out what datasets have dedup''d data in them. Even better would be to discover which files in a particular dataset are dedup''d. I ran # zdb -DDDD which gave output like: index 1055c9f21af63 refcnt 2 single DVA[0]=<0:1e274ec3000:2ac00:STD:1> [L0 deduplicated block] sha256 uncompressed LE contiguous unique unencrypted 1-copy size=20000L/20000P
2011 Nov 25
1
Recovering from kernel panic / reboot cycle importing pool.
Yesterday morning I awoke to alerts from my SAN that one of my OS disks was faulty, FMA said it was in hardware failure. By the time I got to work (1.5 hours after the email) ALL of my pools were in a degraded state, and "tank" my primary pool had kicked in two hot spares because it was so discombobulated. ------------------- EMAIL ------------------- List of faulty resources:
2012 Apr 16
2
Any options on crypt+zfs ?
hail, I have a soekris running an atom and 2GB RAM and ZFS using 7 drives, small capacity though, to test and study if I can make my home server this box and this way. It will be a simple server, three users tops. I followed the handbook and made the geli step on the disks: Geom name: label/zfs1.eli State: ACTIVE EncryptionAlgorithm: AES-XTS KeyLength: 128 Crypto: software UsedKey: 0 Flags:
2006 May 19
11
tracking error to file
In my testing, I''ve found the following error: zpool status -v pool: local state: ONLINE status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore the entire pool from backup. see: http://www.sun.com/msg/ZFS-8000-8A scrub: none requested
2011 Aug 24
3
Efficient way to Calculate the squared distances for a set of vectors to a fixed vector
I am pretty new to R. So this may be an easy question for most of you. ? I would like to calculate the squared distances of a large set (let's say 20000) of vectors (let's say dimension of 5) to a fixed vector. ? Say I have a data frame MY_VECTORS with 20000 rows and 5 columns, and one 5x1 vector y. I would like to efficiently calculate the squared distances?between each of the 20000
2002 Feb 07
0
Solaris 2.7 error in file IO (code 11) at main.c(883)
I'm updating to version 2.5.2 on a Solaris 2.7 (and 2.6) system. The rsync process seems to run but finishes with the following: wrote 91830 bytes read 4589600 bytes 105200.67 bytes/sec total size is 127959731 speedup is 27.33 rsync error: error in file IO (code 11) at main.c(883) I've tried the binary downloaded version and recompiled myself from the source. Is the error message an
2010 Mar 05
2
ZFS replication send/receive errors out
My full backup script errorred out the last two times I ran it. I''ve got a full Bash trace of it, so I know exactly what was done. There are a moderate number of snapshots on the zp1 pool, and I''m intending to replicate the whole thing into the backup pool. After housekeeping, I take make a current snapshot on the data pool (zp1). Since this is a new full backup, I then
2006 Nov 02
4
reproducible zfs panic on Solaris 10 06/06
Hi, I am able to reproduce the following panic on a number of Solaris 10 06/06 boxes (Sun Blade 150, V210 and T2000). The script to do this is: #!/bin/sh -x uname -a mkfile 100m /data zpool create tank /data zpool status cd /tank ls -al cp /etc/services . ls -al cd / rm /data zpool status # uncomment the following lines if you want to see the system think # it can still read and write to the
2007 Apr 23
3
ZFS panic caused by an exported zpool??
Apr 23 02:02:21 SERVER144 offline or reservation conflict Apr 23 02:02:21 SERVER144 scsi: [ID 107833 kern.warning] WARNING: /scsi_vhci/disk at g60001fe100118db00009119074440055 (sd82): Apr 23 02:02:21 SERVER144 i/o to invalid geometry Apr 23 02:02:21 SERVER144 scsi: [ID 107833 kern.warning] WARNING: /scsi_vhci/disk at g60001fe100118db00009119074440055 (sd82): Apr 23 02:02:21 SERVER144
2006 Oct 05
0
Crash when doing rm -rf
Not an really good subject, I know but that''s kind of what happend. I''m trying to build an backup-solution server, Windows users using OSCAR (which uses rsync) to sync their files to an folder and when complete takes a snapshot. It has worked before but then I turned on the -R switch to rsync and when I then removed the folder with rm -rf it crashed. I didn''t save what
2007 May 19
1
Powerware Prestige 1000VA
Hi all, I've more-or-less got NUT 2.0.4 working with my Powerware Prestige 1000VA UPS (serial), but unfortunately NUT is flagging LOW_BATTERY continuously, which is a tad embarrassing when power fails ;-) The numbers from upsc blah at localhost seem reasonable: ambient.temperature: 24.5 battery.charge: 99 battery.runtime: 3999 battery.voltage: 69.2 driver.name: bcmxcp
2007 Oct 21
0
libswfdec-gtk/Makefile.am
libswfdec-gtk/Makefile.am | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) New commits: commit ade9662fa6286636bf22007a5c13898a64bd49b1 Author: Benjamin Otte <otte at gnome.org> Date: Sun Oct 21 16:52:57 2007 +0200 We cann g_thread_init() here, so require gthread diff --git a/libswfdec-gtk/Makefile.am b/libswfdec-gtk/Makefile.am index 5aa9af5..b04b40f 100644 ---
2008 Mar 20
7
ZFS panics solaris while switching a volume to read-only
Hi, I just found out that ZFS triggers a kernel-panic while switching a mounted volume into read-only mode: The system is attached to a Symmetrix, all zfs-io goes through Powerpath: I ran some io-intensive stuff on /tank/foo and switched the device into read-only mode at the same time (symrdf -g bar failover -establish). ZFS went ''bam'' and triggered a Panic: WARNING: /pci at
2007 Nov 09
3
Major problem with a new ZFS setup
We recently installed a 24 disk SATA array with an LSI controller attached to a box running Solaris X86 10 Release 4. The drives were set up in one big pool with raidz, and it worked great for about a month. On the 4th, we had the system kernel panic and crash, and it''s now behaving very badly. Here''s what diagnostic data I''ve been able to collect so far: In the
2013 May 09
0
Memory reservation for 32bit guests on high RAM systems
We''ve been looking in to support of 32bit domains on machines with large amounts of RAM. Working from the ballooning and reservation support added a few years ago: http://lists.xen.org/archives/html/xen-devel/2009-11/msg01277.html 32 bit domains eventually fail to boot with the following Xen complaint: xm create: Error: (4, ''Out of memory'', "panic:
2010 May 28
0
zpool iostat question
Following is the output of "zpool iostat -v". My question is regarding the datapool row and the raidz2 row statistics. The datapool row statistic "write bandwidth" is 381 which I assume takes into account all the disks - although it doesn''t look like it''s an average. The raidz2 row static "write bandwidth" is 36, which is where I am confused. What