similar to: zpool scrub results in pool deadlock

Displaying 20 results from an estimated 10000 matches similar to: "zpool scrub results in pool deadlock"

2007 Sep 08
1
zpool degraded status after resilver completed
I am curious why zpool status reports a pool to be in the DEGRADED state after a drive in a raidz2 vdev has been successfully replaced. In this particular case drive c0t6d0 was failing so I ran, zpool offline home/c0t6d0 zpool replace home c0t6d0 c8t1d0 and after the resilvering finished the pool reports a degraded state. Hopefully this is incorrect. At this point is the vdev in question now has
2007 Apr 02
0
zpool scrub checksum error
Hello, i?ve already read many posts about checksum error on zpools but i like to have some more informations, please. We use 2 sun servers (amd x64, SunOS, 5.10 Generic_118855-36, hopefully all patches) with two hardware raids (raid 10) connected through fibre channel. Disk space is about 3 TB split into 4 pool including several (about 10 - 15) zfs each. After 22 days uptime i got a first
2007 Dec 12
0
Degraded zpool won''t online disk device, instead resilvers spare
I''ve got a zpool that has 4 raidz2 vdevs each with 4 disks (750GB), plus 4 spares. At one point 2 disks failed (in different vdevs). The message in /var/adm/messages for the disks were ''device busy too long''. Then SMF printed this message: Nov 23 04:23:51 x.x.com EVENT-TIME: Fri Nov 23 04:23:51 EST 2007 Nov 23 04:23:51 x.x.com PLATFORM: Sun Fire X4200 M2, CSN:
2010 Aug 19
0
Zpool scrub and reboot.
Suppose I start a zpool scrub and reboot before the scrub is finished. On reboot does the scrub carry on from where it left off or does it start at the beginning again? Thanks -- This message posted from opensolaris.org
2011 Jul 12
1
Can zpool permanent errors fixed by scrub?
Hi, we had a server that lost connection to fiber attached disk array where data luns were housed, due to 3510 power fault. After connection restored alot of the zpool status had these permanent errors listed as per below. I check the files in question and as far as I could see they were present and ok. I ran a zpool scrub against other zpools and they came back with no errors and the list of
2009 Feb 18
4
Zpool scrub in cron hangs u3/u4 server, stumps tech support.
I''ve got a server that freezes when I run a zpool scrub from cron. Zpool scrub runs fine from the command line, no errors. The freeze happens within 30 seconds of the zpool scrub happening. The one core dump I succeeded in taking showed an arccache eating up all the ram. The server''s running Solaris 10 u3, kernel patch 127727-11 but it''s been patched and seems to have
2008 Jul 28
1
zpool status my_pool , shows a pulled disk c1t6d0 as ONLINE ???
New server build with Solaris-10 u5/08, on a SunFire t5220, and this is our first rollout of ZFS and Zpools. Have 8 disks, boot disk is hardware mirrored (c1t0d0 + c1t1d0) Created Zpool my_pool as RaidZ using 5 disks + 1 spare: c1t2d0, c1t3d0, c1t4d0, c1t5d0, c1t6d0, and spare c1t7d0 I am working on alerting & recovery plans for disks failures in the zpool. As a test, I have pulled disk
2011 May 13
27
Extremely slow zpool scrub performance
Running a zpool scrub on our production pool is showing a scrub rate of about 400K/s. (When this pool was first set up we saw rates in the MB/s range during a scrub). Both zpool iostat and an iostat -Xn show lots of idle disk times, no above average service times, no abnormally high busy percentages. Load on the box is .59. 8 x 3GHz, 32GB ram, 96 spindles arranged into raidz zdevs on OI 147.
2007 Apr 27
2
Scrubbing a zpool built on LUNs
I''m building a system with two Apple RAIDs attached. I have hardware RAID5 configured so no RAIDZ or RAIDZ2, just a basic zpool pointing at the four LUNs representing the four RAID controllers. For on-going maintenance, will a zpool scrub be of any benefit? From what I''ve read with this layer of abstration ZFS is only maintaining the metadata and not the actual data on the
2010 Jan 08
0
ZFS partially hangs when removing an rpool mirrored disk while having some IO on another pool on another partition of the same disk
Hello, Sorry for the (very) long subject but I''ve pinpointed the problem to this exact situation. I know about the other threads related to hangs, but in my case there was no < zfs destroy > involved, nor any compression or deduplication. To make a long story short, when - a disk contains 2 partitions (p1=32GB, p2=1800 GB) and - p1 is used as part of a zfs mirror of rpool
2007 Mar 05
3
How to interrupt a zpool scrub?
Dear all Is there a way to stop a running scrub on a zfs pool? Same question applies to a running resilver. Both render our fileserver unusable due to massive CPU load so we''d like to postpone them. In the docs it says that resilvering and scrubbing survive a reboot, so I am not even sure if a reboot would stop scrubbing or resilvering. Any help greatly appreciated! Cheers, Thomas
2010 Aug 30
5
pool died during scrub
I have a bunch of sol10U8 boxes with ZFS pools, most all raidz2 8-disk stripe. They''re all supermicro-based with retail LSI cards. I''ve noticed a tendency for things to go a little bonkers during the weekly scrub (they all scrub over the weekend), and that''s when I''ll lose a disk here and there. OK, fine, that''s sort of the point, and they''re
2005 Dec 22
2
zpool iostat output gets buffered
I''m trying to write a SLAMD (http://www.slamd.com/) resource monitor that can be used to measure the I/O throughput on a ZFS pool, and in particular to be able to get the read and write rates. In order to do this, I''m basically executing "zpool iostat {interval}" and parsing the output to capture the values in the "bandwidth read" and "bandwidth
2008 Aug 25
5
Unable to import zpool since system hang during zfs destroy
Hi all, I have a RAID-Z zpool made up of 4 x SATA drives running on Nexenta 1.0.1 (OpenSolaris b85 kernel). It has on it some ZFS filesystems and few volumes that are shared to various windows boxes over iSCSI. On one particular iSCSI volume, I discovered that I had mistakenly deleted some files from the FAT32 partition that is on it. The files were still in a ZFS snapshot that was made earlier
2007 Oct 14
1
odd behavior from zpool replace.
i''ve got a little zpool with a naughty raidz vdev that won''t take a replacement that as far as i can tell should be adequate. a history: this could well be some bizarro edge case, as the pool doesn''t have the cleanest lineage. initial creation happened on NexentaCP inside vmware in linux. i had given the virtual machine raw device access to 4 500gb drives and 1 ~200gb
2009 Jan 30
1
RFE: parsable iostat and zpool layout
I would like zpool iostat to take a "-p" option to output parsable statistics with absolute counters/figures that for example could be fed to MRTG, RRD, et al. The "zpool iostat [-v] POOL 60 [N]" is great for humans but not very api-friendly; N=2 is a bit overkill and unreliable. Is this info available in kstat, or is this an RFE candidate? In Solaris10? Ditto for zpool
2010 Jun 16
0
files lost in the zpool - retrieval possible ?
Greetings, my Opensolaris 06/2009 installation on an Thinkpad x60 notebook is a little unstable. From the symptoms during installation it seems that there might be something with the ahci driver. No problem with the Opensolaris LiveCD system. Some weeks ago during copy of about 2 GB from a USB stick to the zfs filesystem, the system froze and afterwards refused to boot. Now when investigating
2009 Oct 30
1
internal scrub keeps restarting resilvering?
After several days of trying to get a 1.5TB drive to resilver and it continually restarting, I eliminated all of the snapshot-taking facilities which were enabled and 2009-10-29.14:58:41 [internal pool scrub done txg:567780] complete=0 2009-10-29.14:58:41 [internal pool scrub txg:567780] func=1 mintxg=3 maxtxg=567354 2009-10-29.16:52:53 [internal pool scrub done txg:567999] complete=0
2008 Feb 25
3
[Bug 631] New: zpool get with no pool name dumps core in zfs-crypto
http://defect.opensolaris.org/bz/show_bug.cgi?id=631 Summary: zpool get with no pool name dumps core in zfs-crypto Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: minor Priority: P4 Component: other AssignedTo:
2011 Feb 05
0
40MB repaired on a disk during scrub but no errors
Hey folks, While scrubbing, zpool status shows nearly 40MB "repaired" but 0 in each of the read/write/checksum columns for each disk. One disk has "(repairing)" to the right but once the scrub completes there''s no mention that anything ever needed fixing. Any idea what would need to be repaired on that disk? Are there any other types of errors besides