similar to: Crash when doing rm -rf

Displaying 17 results from an estimated 17 matches similar to: "Crash when doing rm -rf"

2005 Nov 19
11
ZFS related panic!
> My current zfs setup lookst like this: > > homepool 3.63G 34.1G 8K /homepool > > homepool/db 61.6M 34.1G 8.50K /var/db > > homepool/db/pgsql 61.5M 34.1G 61.5M > > /var/db/pgsql > > homepool/home 3.57G 34.1G 10.0K /users > > homepool/home/carrie 8K 34.1G 8K > > /users/carrie > >
2007 Oct 12
0
zfs: allocating allocated segment(offset=77984887808 size=66560)
how does one free segment(offset=77984887808 size=66560) on a pool that won''t import? looks like I found http://bugs.opensolaris.org/view_bug.do?bug_id=6580715 http://mail.opensolaris.org/pipermail/zfs-discuss/2007-September/042541.html when I luupgrade a ufs partition with a dvd-b62 that was bfu to b68 with a dvd of b74 it booted fine and I was doing the same thing that I had done on
2007 Sep 19
3
ZFS panic when trying to import pool
I have a raid-z zfs filesystem with 3 disks. The disk was starting have read and write errors. The disks was so bad that I started to have trans_err. The server lock up and the server was reset. Then now when trying to import the pool the system panic. I installed the last Recommend on my Solaris U3 and also install the last Kernel patch (120011-14). But still when trying to do zpool import
2007 Sep 18
5
ZFS panic in space_map.c line 125
One of our Solaris 10 update 3 servers paniced today with the following error: Sep 18 00:34:53 m2000ef savecore: [ID 570001 auth.error] reboot after panic: assertion failed: ss != NULL, file: ../../common/fs/zfs/space_map.c, line: 125 The server saved a core file, and the resulting backtrace is listed below: $ mdb unix.0 vmcore.0 > $c vpanic() 0xfffffffffb9b49f3() space_map_remove+0x239()
2007 Apr 23
3
ZFS panic caused by an exported zpool??
Apr 23 02:02:21 SERVER144 offline or reservation conflict Apr 23 02:02:21 SERVER144 scsi: [ID 107833 kern.warning] WARNING: /scsi_vhci/disk at g60001fe100118db00009119074440055 (sd82): Apr 23 02:02:21 SERVER144 i/o to invalid geometry Apr 23 02:02:21 SERVER144 scsi: [ID 107833 kern.warning] WARNING: /scsi_vhci/disk at g60001fe100118db00009119074440055 (sd82): Apr 23 02:02:21 SERVER144
2007 Nov 25
2
Corrupted pool
Howdy, We are using ZFS on one of our Solaris 10 servers, and the box paniced this evening with the following stack trace: Nov 24 04:03:35 foo unix: [ID 100000 kern.notice] Nov 24 04:03:35 foo genunix: [ID 802836 kern.notice] fffffe80004a14d0 fffffffffb9b49f3 () Nov 24 04:03:35 foo genunix: [ID 655072 kern.notice] fffffe80004a1550 zfs:space_map_remove+239 () Nov 24 04:03:35 foo genunix: [ID
2006 May 09
3
Possible corruption after disk hiccups...
I''m not sure exactly what happened with my box here, but something caused a hiccup on multiple sata disks... May 9 16:40:33 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci10de,5c at 9/pci-ide at a/ide at 0 (ata6): May 9 16:47:43 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci-ide at 7/ide at 1 (ata3): May 9 16:47:43 sol timeout: abort request, target=0
2006 May 12
1
zfs panic when unpacking open solaris source
When unpacking the solaris source onto a local disk on a system running build 39 I got the following panic: panic[cpu0]/thread=d2c8ade0: really out of space d2c8a7b4 zfs:zio_write_allocate_gang_members+3e6 (e4385ac0) d2c8a7d0 zfs:zio_dva_allocate+81 (e4385ac0) d2c8a7e8 zfs:zio_next_stage+66 (e4385ac0) d2c8a800 zfs:zio_checksum_generate+5e (e4385ac0) d2c8a81c zfs:zio_next_stage+66 (e4385ac0)
2007 Sep 14
9
Possible ZFS Bug - Causes OpenSolaris Crash
I?d like to report the ZFS related crash/bug described below. How do I go about reporting the crash and what additional information is needed? I?m using my own very simple test app that creates numerous directories and files of randomly generated data. I have run the test app on two machines, both 64 bit. OpenSolaris crashes a few minutes after starting my test app. The crash has occurred on
2007 Mar 21
4
HELP!! I can''t mount my zpool!!
Hi all. One of our server had a panic and now can''t mount the zpool anymore! Here is what I get at boot: Mar 21 11:09:17 SERVER142 ^Mpanic[cpu1]/thread=ffffffff90878200: Mar 21 11:09:17 SERVER142 genunix: [ID 603766 kern.notice] assertion failed: ss->ss_start <= start (0x670000b800 <= 0x67 00009000), file: ../../common/fs/zfs/space_map.c, line: 126 Mar 21 11:09:17 SERVER142
2007 Apr 03
2
ZFS panics with dmu_buf_hold_array
Hi, I have been wrestling with ZFS issues since yesterday when one of my disks sort of died. After much wrestling with "zpool replace" I managed to get the new disk in and got the pool to resilver, but since then I have one error left that I can''t clear: pool: data state: ONLINE status: One or more devices has experienced an error resulting in data corruption.
2006 Jun 13
4
ZFS panic while mounting lofi device?
I believe ZFS is causing a panic whenever I attempt to mount an iso image (SXCR build 39) that happens to reside on a ZFS file system. The problem is 100% reproducible. I''m quite new to OpenSolaris, so I may be incorrect in saying it''s ZFS'' fault. Also, let me know if you need any additional information or debug output to help diagnose things. Config: [b]bash-3.00#
2009 Apr 21
0
opensolaris crash in vn_rele()
My newly upgraded opensolaris 2008.11 laptop crashed last weekend. (The OS was installed from the os 2008.11 live-cd and then upgraded using the package manager to snv_111.) I was trying to copy a large virtual pc image from my wife''s imac to the laptop. On a whim I had decided to create a separate zvol in the root pool to contain the image, figuring I could create a vbox with linux or
2008 Jan 17
9
ATA UDMA data parity error
Hey all, I''m not sure if this is a ZFS bug or a hardware issue I''m having - any pointers would be great! Following contents include: - high-level info about my system - my first thought to debugging this - stack trace - format output - zpool status output - dmesg output High-Level Info About My System --------------------------------------------- - fresh
2009 Aug 12
4
zpool import -f rpool hangs
I had the rpool with two sata disks in the miror. Solaris 10 5.10 Generic_141415-08 i86pc i386 i86pc Unfortunately the first disk with grub loader has failed with unrecoverable block write/read errors. Now I have the problem to import rpool after the first disk has failed. So I decided to do: "zpool import -f rpool" only with second disk, but it''s hangs and the system is
2010 Mar 07
2
vectorizing ANOVA over a vectorized linear model
Is it possible to vectorize anova over the output of a vectorized lm? I have a gene expression matrix with each row being a gene and columns for samples. There are several factors with interactions. I can get p values by looping over the matrix with lm and anova, but I would like to make this as computationally efficient as possible. I am able to vectorize the lm command, but when I try to use
2007 May 02
41
gzip compression throttles system?
I just had a quick play with gzip compression on a filesystem and the result was the machine grinding to a halt while copying some large (.wav) files to it from another filesystem in the same pool. The system became very unresponsive, taking several seconds to echo keystrokes. The box is a maxed out AMD QuadFX, so it should have plenty of grunt for this. Comments? Ian