I''ve installed Open Solaris 2009.06 on a machine with 5 identical 1TB
wd green drives to create a ZFS nas. The intended install is one drive
dedicated to the OS and the remaining 4 drives in a raidz1 configuration. The
install is working fine, but creating the raidz1 pool and rebooting is causing
the machine report "Cannot find active partition" upon reboot. Below
is command line output from the live cd during the install and post install.
Note that I''ve tried this a few times, so you''ll see tank
faulted in the post install from previous install attempts.
>> Live CD Install
AVAILABLE DISK SELECTIONS:
0. c7d0 <WDC WD10- WD-WMAV5012875-0001-931.51GB>
/pci at 0,0/pci-ide at 1f,2/ide at 0/cmdk at 0,0
1. c7d1 <WDC WD10- WD-WMAV5011699-0001-931.51GB>
/pci at 0,0/pci-ide at 1f,2/ide at 0/cmdk at 1,0
2. c8d0 <WDC WD10- WD-WMAV5011404-0001-931.51GB>
/pci at 0,0/pci-ide at 1f,2/ide at 1/cmdk at 0,0
3. c8d1 <DEFAULT cyl 60797 alt 2 hd 255 sec 126>
/pci at 0,0/pci-ide at 1f,2/ide at 1/cmdk at 1,0
4. c9d0 <DEFAULT cyl 60797 alt 2 hd 255 sec 126>
/pci at 0,0/pci-ide at 1f,5/ide at 0/cmdk at 0,0
** I Install on c8d1 because there was a previous install on this machine and it
is currently the bootable drive.
>> Post Install
root at nas:~# format
Searching for disks...done
AVAILABLE DISK SELECTIONS:
0. c3d0 <WDC WD10- WD-WMAV5011404-0001-931.51GB>
/pci at 0,0/pci-ide at 1f,2/ide at 1/cmdk at 0,0
1. c3d1 <DEFAULT cyl 60797 alt 2 hd 255 sec 126>
/pci at 0,0/pci-ide at 1f,2/ide at 1/cmdk at 1,0
2. c4d0 <DEFAULT cyl 60797 alt 2 hd 255 sec 126>
/pci at 0,0/pci-ide at 1f,5/ide at 0/cmdk at 0,0
3. c6d0 <WDC WD10- WD-WMAV5012875-0001-931.51GB>
/pci at 0,0/pci-ide at 1f,2/ide at 0/cmdk at 0,0
4. c6d1 <WDC WD10- WD-WMAV5011699-0001-931.51GB>
/pci at 0,0/pci-ide at 1f,2/ide at 0/cmdk at 1,0
root at nas:~# zpool status
pool: rpool
state: ONLINE
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 0
c4d0s0 ONLINE 0 0 0
errors: No known data errors
pool: tank
state: UNAVAIL
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
tank UNAVAIL 0 0 0 insufficient replicas
raidz1 UNAVAIL 0 0 0 corrupted data
c3d0 ONLINE 0 0 0
c3d1 ONLINE 0 0 0
c6d0 ONLINE 0 0 0
c6d1 ONLINE 0 0 0
root at nas:~# zpool destroy tank
root at nas:~# zpool status
pool: rpool
state: ONLINE
scrub: none requested
config:
NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 0
c4d0s0 ONLINE 0 0 0
** I run these commands to show the status of each drive from ZFS''s
perspective.
root at nas:~# zpool create test c3d0
invalid vdev specification
use ''-f'' to override the following errors:
/dev/dsk/c3d0s0 is part of exported or potentially active ZFS pool
tank. Please see zpool(1M).
root at nas:~# zpool create test c3d1
invalid vdev specification
use ''-f'' to override the following errors:
/dev/dsk/c3d1s0 is part of exported or potentially active ZFS pool
rpool. Please see zpool(1M).
root at nas:~# zpool create test c6d0
invalid vdev specification
use ''-f'' to override the following errors:
/dev/dsk/c6d0s0 is part of exported or potentially active ZFS pool
tank. Please see zpool(1M).
root at nas:~# zpool create test c6d1
invalid vdev specification
use ''-f'' to override the following errors:
/dev/dsk/c6d1s0 is part of exported or potentially active ZFS pool
tank. Please see zpool(1M).
root at nas:~# zpool create test c4d0
invalid vdev specification
use ''-f'' to override the following errors:
/dev/dsk/c4d0s0 is part of active ZFS pool rpool. Please see zpool(1M).
If I run `zpool create -f tank raidz1 c3d0 c3d1 c6d0 c6d1` it causes the OS not
to boot saying "Cannot find active partition". If I leave c3d1 out..
ie. `zpool create -f tank raidz1 c3d0 c6d0 c6d1` and reboot everything is fine.
This makes no sense to me since c4d0 is showing up in the zpool status as the OS
drive. Am I missing something? I''ve read through documentation but
can''t seem to find the right piece of documentation to shed light on
what I''m missing.
Thanks in advance.
--
This message posted from opensolaris.org