similar to: ZFS space_map

Displaying 20 results from an estimated 20000 matches similar to: "ZFS space_map"

2007 Feb 11
0
unable to mount legacy vol - panic in zfs:space_map_remove - zdb crashes
I have a 100gb SAN lun in a pool, been running ok for about 6 months. panicked the system this morning. system was running S10U2. In the course of troubleshooting I''ve installed the latest recommended bundle including kjp 118833-36 and zfs patch 124204-03 created as: zpool create zfspool01 /dev/dsk/emcpower0c zfs create zfspool01/nb60openv zfs set mountpoint=legacy zfspool01/nb60openv
2007 Nov 14
0
space_map.c ''ss == NULL'' panic strikes back.
Hi. Someone currently reported a ''ss == NULL'' panic in space_map.c/space_map_add() on FreeBSD''s version of ZFS. I found that this problem was previously reported on Solaris and is already fixed. I verified it and FreeBSD''s version have this fix in place...
2007 Sep 18
5
ZFS panic in space_map.c line 125
One of our Solaris 10 update 3 servers paniced today with the following error: Sep 18 00:34:53 m2000ef savecore: [ID 570001 auth.error] reboot after panic: assertion failed: ss != NULL, file: ../../common/fs/zfs/space_map.c, line: 125 The server saved a core file, and the resulting backtrace is listed below: $ mdb unix.0 vmcore.0 > $c vpanic() 0xfffffffffb9b49f3() space_map_remove+0x239()
2010 Jul 24
2
Severe ZFS corruption, help needed.
I''m running FreeBSD 8.1 with ZFS v15. Recently some time after moving my mirrored pool from one device to another system crashes. From that time on zpool cannot be used/imported - any attempt fails with: solaris assert: sm->space + size <= sm->size, file: /usr/src/sys/moules/zfs/../../cddl/contrib/opensolaris/uts/common/fs/zfs/space_map.c, line: 93 Debugging reveals that:
2010 Sep 18
6
space_map again nuked!!
I''m really angry against ZFS: My server no more reboots because the ZFS spacemap is again corrupt. I just replaced the whole spacemap by recreating a new zpool from scratch and copying back the data with "zfs send & zfs receive". Did it copied corrupt spacemap?! For me its now terminated. I loss to much time and money with this experimental filesystem. My version is Zpool
2008 May 26
2
indiana as nfs server: crash due to zfs
hello all, i have indiana freshly installed on a sun ultra 20 machine. It only does nfs server. During one night, the kernel had crashed, and i got this messages: " May 22 02:18:57 ultra20 unix: [ID 836849 kern.notice] May 22 02:18:57 ultra20 ^Mpanic[cpu0]/thread=ffffff0003d06c80: May 22 02:18:57 ultra20 genunix: [ID 603766 kern.notice] assertion failed: sm->sm_space == 0 (0x40000000 ==
2006 Oct 31
0
6410698 ZFS metadata needs to be more highly replicated (ditto blocks)
Author: billm Repository: /hg/zfs-crypto/gate Revision: 33640e100342f4a847c599f1a1671dda6faf4e05 Log message: 6410698 ZFS metadata needs to be more highly replicated (ditto blocks) 6410700 zdb should support reading raw blocks out of storage pool 6410709 ztest: spa config can change before pool export Files: update: usr/src/cmd/mdb/common/modules/zfs/zfs.c update: usr/src/cmd/zdb/zdb.c update:
2008 Dec 15
15
Need Help Invalidating Uberblock
I have a ZFS pool that has been corrupted. The pool contains a single device which was actually a file on UFS. The machine was accidentally halted and now the pool is corrupt. There are (of course) no backups and I''ve been asked to recover the pool. The system panics when trying to do anything with the pool. root@:/$ zpool status panic[cpu1]/thread=fffffe8000758c80: assertion failed:
2008 Jan 15
1
ZFS-DMU benchmark on Linux
Hi all, You will find in attachment to this mail some benchmarks we made on Linux with pios over ZFS-DMU. There are some interesting things about ZFS tuning and some ideas for breaking a bottleneck we identified in DMU. Thomas LEIBOVICI CEA/DAM - Ile de France -------------- next part -------------- A non-text attachment was scrubbed... Name: ZFS_benchs_Linux.pdf Type: application/pdf Size:
2007 Apr 03
2
ZFS panics with dmu_buf_hold_array
Hi, I have been wrestling with ZFS issues since yesterday when one of my disks sort of died. After much wrestling with "zpool replace" I managed to get the new disk in and got the pool to resilver, but since then I have one error left that I can''t clear: pool: data state: ONLINE status: One or more devices has experienced an error resulting in data corruption.
2007 Sep 19
3
ZFS panic when trying to import pool
I have a raid-z zfs filesystem with 3 disks. The disk was starting have read and write errors. The disks was so bad that I started to have trans_err. The server lock up and the server was reset. Then now when trying to import the pool the system panic. I installed the last Recommend on my Solaris U3 and also install the last Kernel patch (120011-14). But still when trying to do zpool import
2009 Mar 03
8
zfs list extentions related to pNFS
Hi, I am soliciting input from the ZFS engineers and/or ZFS users on an extension to "zfs list". Thanks in advance for your feedback. Quick Background: The pNFS project (http://opensolaris.org/os/project/nfsv41/) is adding a new DMU object set type which is used on the pNFS data server to store pNFS stripe DMU objects. A pNFS dataset gets created with the "zfs
2006 Jan 04
8
Using same ZFS under different kernel versions
I build two zfs filesystems using b29 (from brandz). I then re-installed solaris express b28, preserving the zfs filesystems. When I tried to "zpool import" my zfs filesystems I got a kernel panic: > debugging crash dump vmcore.0 (32-bit) from blackbird > operating system: 5.11 snv_28 (i86pc) > panic message: > ZFS: bad checksum (read on /dev/dsk/c1d0p0 off 24d5e000: zio
2007 Feb 06
4
The ZFS MOS and how DNODES are stored
ZFS documentation lists snapshot limits on any single file system in a pool at 2**48 snaps, and that seems to logically imply that a snap on a file system does not require an update to the pool?s currently active uberblock. That is to say, that if we take a snapshot of a file system in a pool, and then make any changes to that file system, the copy on write behavior induced by the changes will
2008 Jun 26
3
[Bug 2334] New: zpool destroy panics after zfs_force_umount_stress
http://defect.opensolaris.org/bz/show_bug.cgi?id=2334 Summary: zpool destroy panics after zfs_force_umount_stress Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: major Priority: P2 Component: other AssignedTo:
2007 Nov 09
3
Major problem with a new ZFS setup
We recently installed a 24 disk SATA array with an LSI controller attached to a box running Solaris X86 10 Release 4. The drives were set up in one big pool with raidz, and it worked great for about a month. On the 4th, we had the system kernel panic and crash, and it''s now behaving very badly. Here''s what diagnostic data I''ve been able to collect so far: In the
2007 Aug 02
3
ZFS, ZIL, vq_max_pending and OSCON
The slides from my ZFS presentation at OSCON (as well as some additional information) are available at http://www.meangrape.com/ 2007/08/oscon-zfs/ Jay Edwards jay at meangrape.com http://www.meangrape.com -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20070802/f2fa7b08/attachment.html>
2008 Mar 20
7
ZFS panics solaris while switching a volume to read-only
Hi, I just found out that ZFS triggers a kernel-panic while switching a mounted volume into read-only mode: The system is attached to a Symmetrix, all zfs-io goes through Powerpath: I ran some io-intensive stuff on /tank/foo and switched the device into read-only mode at the same time (symrdf -g bar failover -establish). ZFS went ''bam'' and triggered a Panic: WARNING: /pci at
2007 Feb 14
5
Need help making lsof work with ZFS
I contacted the author of ''lsof'' regarding the missing ZFS support. The command works but fails to display any files that are opened by the process in a ZFS filesystem. He indicates that the required ZFS kernel structure definitions (header files) are not shipped with the OS. He further indicated that he rummaged through the OpenSolaris source tree and the files doesn''t
2011 Nov 25
1
Recovering from kernel panic / reboot cycle importing pool.
Yesterday morning I awoke to alerts from my SAN that one of my OS disks was faulty, FMA said it was in hardware failure. By the time I got to work (1.5 hours after the email) ALL of my pools were in a degraded state, and "tank" my primary pool had kicked in two hot spares because it was so discombobulated. ------------------- EMAIL ------------------- List of faulty resources: