Displaying 20 results from an estimated 1000 matches similar to: "ZFS panic caused by an exported zpool??"
2007 Sep 19
3
ZFS panic when trying to import pool
I have a raid-z zfs filesystem with 3 disks. The disk was starting have read and write errors.
The disks was so bad that I started to have trans_err. The server lock up and the server was reset. Then now when trying to import the pool the system panic.
I installed the last Recommend on my Solaris U3 and also install the last Kernel patch (120011-14).
But still when trying to do zpool import
2007 Sep 14
9
Possible ZFS Bug - Causes OpenSolaris Crash
I?d like to report the ZFS related crash/bug described below. How do I go about reporting the crash and what additional information is needed?
I?m using my own very simple test app that creates numerous directories and files of randomly generated data. I have run the test app on two machines, both 64 bit.
OpenSolaris crashes a few minutes after starting my test app. The crash has occurred on
2007 Mar 21
4
HELP!! I can''t mount my zpool!!
Hi all.
One of our server had a panic and now can''t mount the zpool anymore!
Here is what I get at boot:
Mar 21 11:09:17 SERVER142 ^Mpanic[cpu1]/thread=ffffffff90878200:
Mar 21 11:09:17 SERVER142 genunix: [ID 603766 kern.notice] assertion failed: ss->ss_start <= start (0x670000b800 <= 0x67
00009000), file: ../../common/fs/zfs/space_map.c, line: 126
Mar 21 11:09:17 SERVER142
2006 May 09
3
Possible corruption after disk hiccups...
I''m not sure exactly what happened with my box here, but something caused a hiccup on multiple sata disks...
May 9 16:40:33 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci10de,5c at 9/pci-ide at a/ide at 0 (ata6):
May 9 16:47:43 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci-ide at 7/ide at 1 (ata3):
May 9 16:47:43 sol timeout: abort request, target=0
2006 Oct 05
0
Crash when doing rm -rf
Not an really good subject, I know but that''s kind of what happend.
I''m trying to build an backup-solution server, Windows users using OSCAR (which uses rsync) to sync their files to an folder and when complete takes a snapshot. It has worked before but then I turned on the -R switch to rsync and when I then removed the folder with rm -rf it crashed. I didn''t save what
2008 Dec 28
2
zfs mount hangs
Hi,
System: Netra 1405, 4x450Mhz, 4GB RAM and 2x146GB (root pool) and
2x146GB (space pool). snv_98.
After a panic the system hangs on boot and manual attempts to mount
(at least) one dataset in single user mode, hangs.
The Panic:
Dec 27 04:42:11 base ^Mpanic[cpu0]/thread=300021c1a20:
Dec 27 04:42:11 base unix: [ID 521688 kern.notice] [AFT1] errID
0x00167f73.1c737868 UE Error(s)
Dec 27
2008 Nov 13
5
BAD TRAP with Crossbow Beta October 31 2008
Hi.
I tried to send this to the mailing list, but it never showed up in the
archives, so I''m trying the forum instead...
I recently installed the Crossbow Beta October 31 2008 on my
SunFire T1000, and let me first say that I''m very pleased
with the functionality it provides.
What''s not so pleasing is the fact that after installing this,
the computer now get very
2007 Jun 16
5
zpool mirror faulted
I have a strange problem with a faulted zpool (two way mirror):
[root at einstein;0]~# zpool status poolm
pool: poolm
state: FAULTED
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
poolm UNAVAIL 0 0 0 insufficient replicas
mirror UNAVAIL 0 0 0 corrupted data
c2t0d0s0 ONLINE 0
2007 Oct 10
6
server-reboot
Hi.
Just migrated to zfs on opensolaris. I copied data to the server using
rsync and got this message:
Oct 10 17:24:04 zetta ^Mpanic[cpu1]/thread=ffffff0007f1bc80:
Oct 10 17:24:04 zetta genunix: [ID 683410 kern.notice] BAD TRAP:
type=e (#pf Page fault) rp=ffffff0007f1b640 addr=fffffffecd873000
Oct 10 17:24:04 zetta unix: [ID 100000 kern.notice]
Oct 10 17:24:04 zetta unix: [ID 839527 kern.notice]
2008 May 26
2
indiana as nfs server: crash due to zfs
hello all,
i have indiana freshly installed on a sun ultra 20 machine. It only does nfs server. During one night, the kernel had crashed, and i got this messages:
"
May 22 02:18:57 ultra20 unix: [ID 836849 kern.notice]
May 22 02:18:57 ultra20 ^Mpanic[cpu0]/thread=ffffff0003d06c80:
May 22 02:18:57 ultra20 genunix: [ID 603766 kern.notice] assertion failed: sm->sm_space == 0 (0x40000000 ==
2010 Feb 08
5
zfs send/receive : panic and reboot
<copied from opensolaris-dicuss as this probably belongs here.>
I kept on trying to migrate my pool with children (see previous threads) and had the (bad) idea to try the -d option on the receive part.
The system reboots immediately.
Here is the log in /var/adm/messages
Feb 8 16:07:09 amber unix: [ID 836849 kern.notice]
Feb 8 16:07:09 amber ^Mpanic[cpu1]/thread=ffffff014ba86e40:
Feb 8
2007 Nov 25
2
Corrupted pool
Howdy,
We are using ZFS on one of our Solaris 10 servers, and the box paniced
this evening with the following stack trace:
Nov 24 04:03:35 foo unix: [ID 100000 kern.notice]
Nov 24 04:03:35 foo genunix: [ID 802836 kern.notice] fffffe80004a14d0
fffffffffb9b49f3 ()
Nov 24 04:03:35 foo genunix: [ID 655072 kern.notice] fffffe80004a1550
zfs:space_map_remove+239 ()
Nov 24 04:03:35 foo genunix: [ID
2007 Aug 26
3
Kernel panic receiving incremental snapshots
Before I open a new case with Sun, I am wondering if anyone has seen this
kernel panic before? It happened on an X4500 running Sol10U3 while it was
receiving incremental snapshot updates.
Thanks.
Aug 25 17:01:50 ldasdata6 ^Mpanic[cpu0]/thread=fffffe857d53f7a0:
Aug 25 17:01:50 ldasdata6 genunix: [ID 895785 kern.notice] dangling dbufs (dn=fffffe82a3532d10, dbuf=fffffe8b4e338b90)
Aug 25 17:01:50
2012 Apr 17
10
kernel panic during zfs import [UPDATE]
Hello everybody,
just to let you know what happened in the meantime:
I was able to open a Service Request at Oracle.
The issue is a known bug (Bug 6742788 : assertion panic at: zfs:zap_deref_leaf)
The bug has bin fixed (according to Oracle support) since build 164, but there is no fix for Solaris 11 available so far (will be fixed in S11U7?).
There is a workaround available that works
2006 Oct 25
4
Panic while scrubbing
Hello,
I am not sure if I am posting in the correct forum, but it seems somewhat zfs related, so I thought I''d share it.
While the machine was idle, I started a scrub. Around the time the scrubbing was supposed to be finished, the machine panicked.
This might be related to the ''metadata corruption'' that happened earlier to me. Here is the log, any ideas?
Oct 24
2008 Jun 18
1
snv_81 domU: long delay during boot when dom0 has been up for a long time
This one happened today:
Gentoo linux dom0, xen 3.2.0 hypervisor (32-bit), Intel Quad-core cpu, 8GB memory.
This linux box / dom0 is up for 40 days.
OpenSolaris SXCE snv_81 PV domU (32-bit)
Problem: when starting the snv_81 domU, we''re stuck for more than 60 seconds
early during bootstrap:
v3.2.0 chgset ''unavailable''
SunOS Release 5.11 Version snv_81 32-bit
2008 Jun 04
2
panic on `zfs export` with UNAVAIL disks
hi list,
initial situation:
SunOS alusol 5.11 snv_86 i86pc i386 i86pc
SunOS Release 5.11 Version snv_86 64-bit
3 USB HDDs on 1 USB hub:
zpool status:
state: ONLINE
NAME STATE READ WRITE CKSUM
usbpool ONLINE 0 0 0
mirror ONLINE 0 0 0
c7t0d0p0 ONLINE 0 0 0
c8t0d0p0
2008 Apr 24
0
panic on zfs scrub on builds 79 & 86
This just started happening to me. It''s a striped non mirrored pool (I know I know). A zfs scrub causes a panic under a minute. I can also trigger a panic by doing tars etc. x86 64-bit kernel ... any ideas? Just to help rule out some things, I changed the motherboard, memory and cpu and it still happens ... I also think it happens on a 32-bit kernel.
genunix: [ID 335743 kern.notice] BAD
2011 Nov 25
1
Recovering from kernel panic / reboot cycle importing pool.
Yesterday morning I awoke to alerts from my SAN that one of my OS disks was faulty, FMA said it was in hardware failure. By the time I got to work (1.5 hours after the email) ALL of my pools were in a degraded state, and "tank" my primary pool had kicked in two hot spares because it was so discombobulated.
------------------- EMAIL -------------------
List of faulty resources:
2005 Nov 19
11
ZFS related panic!
> My current zfs setup lookst like this:
> > homepool 3.63G 34.1G 8K /homepool
> > homepool/db 61.6M 34.1G 8.50K /var/db
> > homepool/db/pgsql 61.5M 34.1G 61.5M
> > /var/db/pgsql
> > homepool/home 3.57G 34.1G 10.0K /users
> > homepool/home/carrie 8K 34.1G 8K
> > /users/carrie
> >