similar to: internal scrub keeps restarting resilvering?

Displaying 20 results from an estimated 300 matches similar to: "internal scrub keeps restarting resilvering?"

2006 Oct 31
0
6344108 snapshot create/delete interlock with scrub/resilver must sync txg
Author: bonwick Repository: /hg/zfs-crypto/gate Revision: 41acc27e604047650771dceb4535b8586bd34848 Log message: 6344108 snapshot create/delete interlock with scrub/resilver must sync txg Files: update: usr/src/cmd/ztest/ztest.c update: usr/src/uts/common/fs/zfs/spa.c
2012 Jan 07
14
zfs defragmentation via resilvering?
Hello all, I understand that relatively high fragmentation is inherent to ZFS due to its COW and possible intermixing of metadata and data blocks (of which metadata path blocks are likely to expire and get freed relatively quickly). I believe it was sometimes implied on this list that such fragmentation for "static" data can be currently combatted only by zfs send-ing existing
2006 Apr 27
2
summary(lm(x~y)) difference between R-2.2.1 and R-2.3.0
Hi [macOSX 10.4.6; R-2.3.0] I have encountered a difference in behaviour between R-2.2.1 and R-2.3.0 when performing a linear model. Transcript follows for R-2.3.0 (R-2.2.1 worked as expected). How to make R-2.3.0 perform as R-2.2.1 did? > dput(x) c(29.13, 29.88, 30.09, 29.99, 29.74, 29.64, 29.65, 29.7, 30.04, 29.89, 29.96, 29.65, 28.76, 28.41, 28.38, 29.55, 29.76, 29.75, 29.84,
2010 Dec 20
3
Resilvering - Scrubing whats the different
Hello All I read this thread Resilver/scrub times? for a few minutes and I have recognize that I dont know the different between Resilvering and Scrubing. Shame on me. :-( I dont find some declarations in the man-pages and I know the command to start scrubing "zpool scrub tank" but what is the command to start resilver and what is the different? -- Best Regards Alexander Dezember, 20
2006 Nov 30
0
ZFS caught resilvering when only one side of mirror persent
When I booted my laptop up this morning it took much longer than normal and there was a lot of disk activity even after I logged in. A quick use of dtrace and iostat revealed that all the writes were to the zpool. I ran zpool status and found that the pool was resilvering. Strange thing is that while the pool is a mirror, one side of it is offline since it is on an external usb disk - which
2006 May 17
0
OCFS2 with Oracle Resilvering.
Hi, Can anyone tell me if OCFS2 is having any support for Oracle resilvering feature? Does OCFS2 co-ordinate with Oracle resilvering mechanism before starting its recovery in case of crash and so? Thanks. Sumit.
2008 Mar 07
0
Sending notification when zpool replace has finished resilvering?
This might be a daft question, but is there any way to configure Solaris to notify us when a disk has finished resilvering? This message posted from opensolaris.org
2010 Mar 07
0
strangeness after resilvering disk from raidz1 on disks with no EFI GPTs
I have a zpool of five 1.5TB disks in raidz1. They are on c?t?d?p0 devices - using the full disk, not any slice or partition, because the pool was created in zfs-fuse in linux and no partition tables were ever created. (for the full saga of my move from that to opensolaris, anyone who missed out on the fun can read the thread http://www.mail-archive.com/zfs-discuss at opensolaris.org/msg34813.html
2009 Jul 11
5
Resilvering Loop
I have a situation where my zpool (with two radiz2s) is resilvering and reaches a certain point, then starts over. There no read, write or checksum errors. The disks do have a fair amount of resilvering to do, as I''ve had a variety of disk failures. But at the core of things, there''s enough parity data to avoid data loss, and the zpool is functioning fine, albeit in a
2008 Sep 05
0
raidz pool metadata corrupted nexanta-core->freenas 0.7->nexanta-core
I made a bad judgment and now my raidz pool is corrupted. I have a raidz pool running on Opensolaris b85. I wanted to try out freenas 0.7 and tried to add my pool to freenas. After adding the zfs disk, vdev and pool. I decided to back out and went back to opensolaris. Now my raidz pool will not mount and got the following errors. Hope someone expert can help me recover from this error.
2009 Jul 10
5
Slow Resilvering Performance
I know this topic has been discussed many times... but what the hell makes zpool resilvering so slow? I''m running OpenSolaris 2009.06. I have had a large number of problematic disks due to a bad production batch, leading me to resilver quite a few times, progressively replacing each disk as it dies (and now preemptively removing disks.) My complaint is that resilvering ends up
2009 Jun 19
8
x4500 resilvering spare taking forever?
I''ve got a Thumper running snv_57 and a large ZFS pool. I recently noticed a drive throwing some read errors, so I did the right thing and zfs replaced it with a spare. Everything went well, but the resilvering process seems to be taking an eternity: # zpool status pool: bigpool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was
2020 Jul 16
2
read.csv fails in R console in Ubuntu terminal but works in RStudio after R 3.6.3 upgrade to R 4.0.2
On 7/15/20 1:35 PM, Dirk Eddelbuettel wrote: > On 15 July 2020 at 16:16, Sam H wrote: > | I am trying to download some data using read.csv and it works perfectly in > | RStudio and fails in the R console in the terminal in Ubuntu 18.04 after > | upgrading from R 3.6.3 to 4.0.2. Before upgrading this worked in the R > | console in the terminal also without any issues. > | > |
2014 Apr 18
0
[PATCH] nouveau/codegen: add missing values for OP_TXLQ into the target arrays
Also rework things so that if someone were to add an opcode without adjusting the values in these arrays, there will be a compilation error. This fixes a few quadop-related piglit regressions since commit d5faf8e78603. Signed-off-by: Ilia Mirkin <imirkin at alum.mit.edu> --- src/gallium/drivers/nouveau/codegen/nv50_ir_target.cpp | 12 +++++++-----
2010 Jun 29
0
Processes hang in /dev/zvol/dsk/poolname
After multiple power outages caused by storms coming through, I can no longer access /dev/zvol/dsk/poolname, which are hold l2arc and slog devices in another pool I don''t think this is related, since I the pools are ofline pending access to the volumes. I tried running find /dev/zvol/dsk/poolname -type f and here is the stack, hopefully this someone a hint at what the issue is, I have
2012 Jun 18
1
Restore destroyed snapshot ???
OK, I am a butt-head and accidentally destroyed my last snapshot of a replicated ZFS dataset. The dataset is NOT mounted and other than a resilver going on, there is no I/O going on to this dataset. Is there any way to roll back and get my latest snapshot back? from zpool history -i: 2012-06-18.10:34:00 zfs destroy xxx at 1339668001 2012-06-18.10:34:00 [internal destroy txg:2213852] dataset =
2013 Feb 17
13
zfs raid1 error resilvering and mount
hi, i have raid1 on zfs with 2 device on pool first device died and boot from second not working... i try to get http://mfsbsd.vx.sk/ flash and load from it with zpool import http://puu.sh/2402E when i load zfs.ko and opensolaris.ko i see this message: Solaris: WARNING: Can''t open objset for zroot/var/crash Solaris: WARNING: Can''t open objset for zroot/var/crash zpool status:
2020 Jul 15
0
read.csv fails in R console in Ubuntu terminal but works in RStudio after R 3.6.3 upgrade to R 4.0.2
On 15 July 2020 at 16:16, Sam H wrote: | I am trying to download some data using read.csv and it works perfectly in | RStudio and fails in the R console in the terminal in Ubuntu 18.04 after | upgrading from R 3.6.3 to 4.0.2. Before upgrading this worked in the R | console in the terminal also without any issues. | | Why would that be? How to fix this? | | Below please find R code output and
2010 Nov 01
6
Excruciatingly slow resilvering on X4540 (build 134)
Hello, I''m working with someone who replaced a failed 1TB drive (50% utilized), on an X4540 running OS build 134, and I think something must be wrong. Last Tuesday afternoon, zpool status reported: scrub: resilver in progress for 306h0m, 63.87% done, 173h7m to go and a week being 168 hours, that put completion at sometime tomorrow night. However, he just reported zpool status shows:
2007 Sep 18
5
ZFS panic in space_map.c line 125
One of our Solaris 10 update 3 servers paniced today with the following error: Sep 18 00:34:53 m2000ef savecore: [ID 570001 auth.error] reboot after panic: assertion failed: ss != NULL, file: ../../common/fs/zfs/space_map.c, line: 125 The server saved a core file, and the resulting backtrace is listed below: $ mdb unix.0 vmcore.0 > $c vpanic() 0xfffffffffb9b49f3() space_map_remove+0x239()