similar to: Messed up zpool (double device label)

Displaying 20 results from an estimated 120 matches similar to: "Messed up zpool (double device label)"

2010 Apr 24
6
Extremely slow raidz resilvering
Hello everyone, As one of the steps of improving my ZFS home fileserver (snv_134) I wanted to replace a 1TB disk with a newer one of the same vendor/model/size because this new one has 64MB cache vs. 16MB in the previous one. The removed disk will be use for backups, so I thought it''s better off to have a 64MB cache disk in the on-line pool than in the backup set sitting off-line all
2002 Jul 19
1
lilo causes a "Unexpected dirty buffer encountered at do_get_write_access:597 (03:02 blocknr 0)"
On my Debian box: Package: lilo Version: 1:22.2-5 Severity: normal lilo seems to cause a kernel warning (see subject) when / is a ext3 partition. Maybe a kernel problem? Who knows. I'm running 2.4.19-rc1-ac7 -- System Information: Debian Release: testing/unstable Architecture: i386 Kernel: Linux hummus 2.4.19-rc1-ac7 #1 Wed Jul 17 22:14:20 CEST 2002 i686 Locale: LANG=C, LC_CTYPE=C Versions
2001 Aug 14
1
[BUG] linux-2.4.7-ac7 Assertion failure in journal_revoke() at revoke.c:307
Greetings all, I have hit a kernel BUG in revoke.c in kernel 2.4.7-ac7 twice today while attempting to perform the same operation (patching stock 2.4.8 kernel src with "patch -p1 < patch-2.4.8-ac4"). Syslog entries follow. Please email me if you want/need my kernel config or any other information. Thanks, jtp
2010 Apr 14
1
Checksum errors on and after resilver
Hi all, I recently experienced a disk failure on my home server and observed checksum errors while resilvering the pool and on the first scrub after the resilver had completed. Now everything seems fine but I''m posting this to get help with calming my nerves and detect any possible future faults. Lets start with some specs. OSOL 2009.06 Intel SASUC8i (w LSI 1.30IT FW) Gigabyte
2001 Sep 06
1
Changing root journal data mode
Hi, I'm running 2.4.9-ac7 with ext3 compiled in, and I ran across this problem. Now that I look back, it makes sence, but it wasn't obvious at first... I wanted to change the journal mode on my / partition, so I changed my fstab to: /dev/hda1 / ext3 defaults,errors=remount-ro,data=journal 0 1 as well as several other partitions, and rebooted. After rebooting I had a read only / that I
2010 May 05
3
[indiana-discuss] image-update doesn''t work anymore (bootfs not supported on EFI)
On 5/5/10 1:44 AM, Christian Thalinger wrote: > On Tue, 2010-05-04 at 16:19 -0600, Evan Layton wrote: >> Can you try the following and see if it really thinks it''s an EFI lable? >> # dd if=/dev/dsk/c12t0d0s2 of=x skip=512 bs=1 count=10 >> # cat x >> >> This may help us determine if this is another instance of bug 6860320 > > # dd
2008 Jan 17
9
ATA UDMA data parity error
Hey all, I''m not sure if this is a ZFS bug or a hardware issue I''m having - any pointers would be great! Following contents include: - high-level info about my system - my first thought to debugging this - stack trace - format output - zpool status output - dmesg output High-Level Info About My System --------------------------------------------- - fresh
2008 Sep 01
1
Running Warcraft on dedicated X server using Bash Script
I am using the following script that has been suggested by many guides on running World of Warcraft on Linux. Code: #!/bin/sh export WOW_PATH=~/".wine/drive_c/Program Files/World of Warcraft" # Installation path X :3 -ac -terminate & # Launch on a new X session on display 3 cd "${WOW_PATH}" # Goto WoW dir sleep 2 DISPLAY=:3 `which wine` Wow.exe -opengl #
2005 Jun 16
0
Error with smbfs
Hello, with kernel 2.6.11.11 and above (I've tried 2.6.11.12 also), I get error in write on smbfs with Novell NetWare server. New kernel work correctly when I get services from other server (Win2k server or pro) I've tested the Alan-Cox's patch 7 also, but without succesfull. My previos kernel (2.4.24) work correctly. Smbclient work correctly with new kernels. Report error from
2012 Jan 08
0
Pool faulted in a bad way
Hello, I have been asked to take a look at at poll on a old OSOL 2009.06 host. It have been left unattended for a long time and it was found in a FAULTED state. Two of the disks in the raildz2 pool seems to have failed, one have been replaced by a spare, the other one is UNAVAIL. The machine was restarted and the damaged disks was removed to make it possible to access the pool without it hanging
2001 Aug 20
1
[tytso@mit.edu: Re: Your ext2 optimisation for readdir+stat]
I asked Ted about his patch for ext3 and ext2 and this is his response. I am not sure if his post will reach lkml since it's sent to vger.rutgers.edu IMHO, With the amount of work Andrew has done to make ext3 usable for MTA applications, Ted's work would make ext3 even better for MTA apps sinc both Postfix/qmail use 'find' in their control-scripts and queue statistics program
2011 Aug 09
7
Disk IDs and DD
Hiya, Is there any reason (and anything to worry about) if disk target IDs don''t start at 0 (zero). For some reason mine are like this (3 controllers - 1 onboard and 2 PCIe); AVAILABLE DISK SELECTIONS: 0. c8t0d0 <ATA -ST9160314AS -SDM1 cyl 19454 alt 2 hd 255 sec 63> /pci at 0,0/pci10de,cb84 at 5/disk at 0,0 1. c8t1d0 <ATA -ST9160314AS -SDM1
2010 Feb 24
0
disks in zpool gone at the same time
Hi, Yesterday I got all my disks in two zpool disconected. They are not real disks - LUNS from StorageTek 2530 array. What could that be - a failing LSI card or a mpt driver in 2009.06? After reboot got four disks in FAILED state - zpool clear fixed things with resilvering. Here is how it started (/var/adm/messages) Feb 23 12:39:03 nexus scsi: [ID 365881 kern.info] /pci at 0,0/pci10de,5d at
2009 Jan 21
8
cifs perfomance
Hello! I''am setup zfs / cifs home storage server, end now have low performance with play movie stored on this zfs from windows client. server hardware is not new , but n windows it perfomance was normal. CPU is AMD Athlon Burton Thunderbird 2500, runing on 1,7GHz, 1024 RAM and storage: usb c4t0d0 ST332062-0A-3.AA-298.09GB /pci at 0,0/pci1458,5004 at 2,2/cdrom at 1/disk at
2009 Jul 29
0
LVM and ZFS
I''m curious about if there are any potential problems with using LVM metadevices as ZFS zpool targets. I have a couple of situations where using a device directly by ZFS causes errors on the console about "Bus and lots of "stalled" I/O. But as soon as I wrap that device inside an LVM metadevice and then use it in the ZFS zpool things work perfectly fine and smoothly (no
2007 Oct 14
1
odd behavior from zpool replace.
i''ve got a little zpool with a naughty raidz vdev that won''t take a replacement that as far as i can tell should be adequate. a history: this could well be some bizarro edge case, as the pool doesn''t have the cleanest lineage. initial creation happened on NexentaCP inside vmware in linux. i had given the virtual machine raw device access to 4 500gb drives and 1 ~200gb
2008 Sep 05
0
raidz pool metadata corrupted nexanta-core->freenas 0.7->nexanta-core
I made a bad judgment and now my raidz pool is corrupted. I have a raidz pool running on Opensolaris b85. I wanted to try out freenas 0.7 and tried to add my pool to freenas. After adding the zfs disk, vdev and pool. I decided to back out and went back to opensolaris. Now my raidz pool will not mount and got the following errors. Hope someone expert can help me recover from this error.
2001 Aug 12
0
Half-Life, Debian, Sid, Wine, Crap, VooDoo, Banshee, OpenGL, Mesa, Glide
Well. i got s.th. for you and i think i'll give this to crazney (lhl.linuxgames.com) #1 a step by step HL tutorial #2 my questions. -------------------------------------------------------------------------- ## 1 ## a step by step HL tutorial ###################################### -------------------------------------------------------------------------- yea, well, i got HL working. On my
2006 May 09
3
Possible corruption after disk hiccups...
I''m not sure exactly what happened with my box here, but something caused a hiccup on multiple sata disks... May 9 16:40:33 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci10de,5c at 9/pci-ide at a/ide at 0 (ata6): May 9 16:47:43 sol scsi: [ID 107833 kern.warning] WARNING: /pci at 0,0/pci-ide at 7/ide at 1 (ata3): May 9 16:47:43 sol timeout: abort request, target=0
2004 Mar 01
0
failure notice (PR#6630)
Hi. This is the qmail-send program at provida.org.br. I'm afraid I wasn't able to deliver your message to the following addresses. This is a permanent error; I've given up. Sorry it didn't work out. <musical@provida.org.br>: vdeliver: ERRO: Usu?rio n?o existe: 'musical'. --- Below this line is a copy of the message. Return-Path: <r-bugs@r-project.org>