similar to: strange ''too many errors'' msg

Displaying 20 results from an estimated 400 matches similar to: "strange ''too many errors'' msg"

2006 Oct 24
3
determining raidz pool configuration
Hi all, Sorry for the newbie question, but I''ve looked at the docs and haven''t been able to find an answer for this. I''m working with a system where the pool has already been configured and want to determine what the configuration is. I had thought that''d be with zpool status -v <poolname>, but it doesn''t seem to agree with the
2007 Feb 27
16
understanding zfs/thunoer "bottlenecks"?
Currently I''m trying to figure out the best zfs layout for a thumper wrt. to read AND write performance. I did some simple mkfile 512G tests and found out, that per average ~ 500 MB/s seems to be the maximum on can reach (tried initial default setup, all 46 HDDs as R0, etc.). According to http://www.amd.com/us-en/assets/content_type/DownloadableAssets/ArchitectureWP_062806.pdf I would
2008 Apr 02
1
delete old zpool config?
Hi experts zpool import shows some weird config of an old zpool bash-3.00# zpool import pool: data1 id: 7539031628606861598 state: FAULTED status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: http://www.sun.com/msg/ZFS-8000-3C config: data1 UNAVAIL insufficient replicas
2007 Apr 11
0
raidz2 another resilver problem
Hello zfs-discuss, One of a disk started to behave strangely. Apr 11 16:07:42 thumper-9.srv sata: [ID 801593 kern.notice] NOTICE: /pci at 1,0/pci1022,7458 at 3/pci11ab,11ab at 1: Apr 11 16:07:42 thumper-9.srv port 6: device reset Apr 11 16:07:42 thumper-9.srv scsi: [ID 107833 kern.warning] WARNING: /pci at 1,0/pci1022,7458 at 3/pci11ab,11ab at 1/disk at 6,0 (sd27): Apr 11 16:07:42 thumper-9.srv
2008 Mar 12
3
Mixing RAIDZ and RAIDZ2 zvols in the same zpool
I have a customer who has implemented the following layout: As you can see, he has mostly raidz zvols but has one raidz2 in the same zpool. What are the implications here? Is this a bad thing to do? Please elaborate. Thanks, Scott Gaspard Scott.J.Gaspard at Sun.COM > NAME STATE READ WRITE CKSUM > > chipool1 ONLINE 0 0 0 > >
2009 Jun 19
8
x4500 resilvering spare taking forever?
I''ve got a Thumper running snv_57 and a large ZFS pool. I recently noticed a drive throwing some read errors, so I did the right thing and zfs replaced it with a spare. Everything went well, but the resilvering process seems to be taking an eternity: # zpool status pool: bigpool state: ONLINE status: One or more devices has experienced an unrecoverable error. An attempt was
2008 Apr 11
0
How to replace root drive if ZFS data is on it?
Hi, Experts: A customer has X4500 and the boot drives mirrored (c5t0d0s0 and c5t4d0s0) by SVM, The ZFS uses the two other partitions on these two drives(c5t0d0s3 and c5t4d0s3). If we need to replace the disk drive c5t0d0, do we need to do anything on the ZFS (c5t0d0s3 and c5t4d0s3) first or just follow the regular boot drive replacement procedure? Below is the summary of their current ZFS
2010 Feb 20
2
R logo as SVG ?
Hi, does anybody have the R logo in a vector format preferable SVG? Need it for Freedesktop (GNOME desktop) integration of Rcmdr ... Thanx, jel. -- Otto-von-Guericke University http://www.cs.uni-magdeburg.de/ Department of Computer Science Geb. 29 R 027, Universitaetsplatz 2 39106 Magdeburg, Germany Tel: +49 391 67 12768
2008 Oct 12
3
blue screen: clock interrupt win2008
Hi, when installing windows server 2008 Std hvm I get always a blue screen telling me: ''a clock interrupt was not received on a secondary processor within the allocated time interval'' I googled around and found, that this occured with vista as well: http://lists.xensource.com/archives/html/xen-changelog/2007-01/msg00170.html Config is: + virt-install --name=win2008ss
2010 Jul 22
6
Xen EOF?
Hi, just found: PSARC/2010/250 EOF of Solaris xVM dom0 approved. So does this mean, Xen is officially dead on Solaris? Regards, jel. -- Otto-von-Guericke University http://www.cs.uni-magdeburg.de/ Department of Computer Science Geb. 29 R 027, Universitaetsplatz 2 39106 Magdeburg, Germany Tel: +49 391 67 12768
2010 Jan 22
0
Removing large holey file does not free space 6792701 (still)
Hello, I mentioned this problem a year ago here and filed 6792701 and I know it has been discussed since. It should have been fixed in snv_118, but I can still trigger the same problem. This is only triggered if the creation of a large file is aborted, for example by loss of power, crash or SIGINT to mkfile(1M). The bug should probably be reopened but I post it here since some people where
2009 Dec 31
1
readCitationFile encoding
Hi, does anybody know, how to enforce a certain encoding for readCitationFile()? Haven't found any usable documentation for this :( E.g.: basecit <- system.file("CITATION", package="base") source(basecit, echo=TRUE, encoding="8859-1") readCitationFile(basecit) BTW: 'make check' is a relativ long running task and if one changes a source file, all
2008 Jan 17
9
ATA UDMA data parity error
Hey all, I''m not sure if this is a ZFS bug or a hardware issue I''m having - any pointers would be great! Following contents include: - high-level info about my system - my first thought to debugging this - stack trace - format output - zpool status output - dmesg output High-Level Info About My System --------------------------------------------- - fresh
2011 Sep 11
8
bad seagate drive?
Hi list, I''ve got a system with 3 WD and 3 seagate drives. Today I got an email that zpool status indicated one of the seagate drives as REMOVED. I''ve tried clearing the error but the pool becomes faulted again. Taken out the offending drive and plugged into a windows box with seatools install. Unfortunately seatools finds nothing wrong with the drive. Windows seems to see
2010 Dec 05
4
Zfs ignoring spares?
Hi all I have installed a new server with 77 2TB drives in 11 7-drive RAIDz2 VDEVs, all on WD Black drives. Now, it seems two of these drives were bad, one of them had a bunch of errors, the other was very slow. After zfs offlining these and then zfs replacing them with online spares, resilver ended and I thought it''d be ok. Appearently not. Albeit the resilver succeeds, the pool status
2009 Sep 07
4
Upgrade from b113 to b122 - any gotchas
We are running 4 Solaris10u7 DomUs on a b113 Dom0 I''d like to upgrade the Dom0 to b122 but not rebuild the DomUs. Below is a typical DomU xml dump. Is there anything obvious that will bite me in upgrading from 113 to 122? Thanks John virsh # dumpxml achilles2 <domain type=''xen'' id=''1''> <name>achilles2</name>
2008 Nov 12
21
zfs boot - U6 kernel patch breaks sparc boot
Hi, in preparation to try zfs boot on sparc I installed all recent patches incl. feature patches comming from s10s_u3wos_10 and after reboot finally 137137-09 (still having everything on UFS). Now it doesn''t boot at anymore: ############################### Sun Fire V240, No Keyboard Copyright 2006 Sun Microsystems, Inc. All rights reserved. OpenBoot 4.22.23, 2048 MB memory installed,
2008 Oct 16
1
attaching 2nd vol unsupported?
Hi, im trying to attach another volume aka disk to win HVM, however it doesn''t seem to work: + xm block-attach win2008ss phy:/dev/zvol/dsk/pool1/win2008ss.dsk2 \ hdd:disk w 0 results in: elkner.sol ~ > + xm block-list win2008ss --long (0 (vbd (uuid 7cb8fadf-619d-dde6-bda9-dcc18023c7d5) (bootable 1) (devid 768) (driver paravirtualised)
2006 Jul 18
1
file access algorithm within pools
Hello, What is the access algorithm used within multi-component pools for a given pool, and does it change when one or more members of the pool become degraded ? examples: zpool create mtank mirror c1t0d0 c2t0d0 mirror c3t0d0 c4t0d0 mirror c5t0d0 c6t0d0 or; zpool create ztank raidz c1t0d0 c2t0d0 c3t0d0 raidz c4t0d0 c5t0d0 c6t0d0 As files are created on the filesystem within these pools,
2009 Oct 15
10
xVM Ready for serious use ?
I currently have 2 x ESXi boxes that have VMs stored on a NFS/iSCSI debian lenny linux box. I had purchased a new whitebox server as an intended replacement for the linux box. I had always planned on installing opensolaris on the new hardware with this config; 2 x 640GB mirrored rpool 6 x 1TB in raidz2 This box was going to be built in parallel with the current server and I would have simply