IHAC who is having problem in importing zpool. When tried to import zpool, it is failing with error message: # zpool import -f Backup1 cannot import ''Backup1'': invalid vdev configuration fma log reports a bad label on vdev_guid 0xd51633a1766882ad. fma errors: Jan 23 2010 09:18:27.374175886 ereport.fs.zfs.vdev.bad_label nvlist version: 0 class = ereport.fs.zfs.vdev.bad_label ena = 0xa359146fa900c01 detector = (embedded nvlist) nvlist version: 0 version = 0x0 scheme = zfs pool = 0x858ffdc3116bac vdev = 0xd51633a1766882ad (end detector) pool = Backup1 pool_guid = 0x858ffdc3116bac pool_context = 2 pool_failmode = wait vdev_guid = 0xd51633a1766882ad vdev_type = raidz parent_guid = 0x858ffdc3116bac parent_type = root prev_state = 0x7 __ttl = 0x1 __tod = 0x4b5b0533 0x164d788e zpool import reports "insufficient replicas" because first raidz2 group is not available. Comparing it with the older zpool import, I see few drives has the controller number changed and there are duplicate entries for c7t0d0. I want to know how to map fma vdev_guid to the device. Running zdb-l on all the devices in first raidz2, I could not able to find this vdev_guid. Also label on all the drives is pointing to c8t0d0, c8t1d0. What steps should be taken to import zpool in degraded mode. raidz2 can sustain two drives failure. Should I try removing /dev/dsk, /dev/rdsk entries for c7t0d0 and c4t1d0 to see if it works? Is there a better way to approach this issue? [root at sanback1]/># zpool import pool: Backup1 id: 37594491964713900 state: UNAVAIL status: The pool is formatted using an older on-disk version. action: The pool cannot be imported due to damaged devices or data. config: Backup1 UNAVAIL insufficient replicas raidz2 UNAVAIL corrupted data c0t0d0 ONLINE c1t0d0 ONLINE c5t1d0 ONLINE c7t0d0 ONLINE << duplicate c7t0d0 ONLINE << duplicate c4t1d0 ONLINE << it was c8t1d0 c0t1d0 ONLINE c4t2d0 ONLINE << it was c8t0d0 c5t2d0 ONLINE c6t1d0 ONLINE c7t1d0 ONLINE raidz2 ONLINE ..... zpool status output from last successful import shows: root at SANBack1:/# zpool status pool: Backup1 state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM Backup1 ONLINE 0 0 0 raidz2 ONLINE 0 0 0 c0t0d0 ONLINE 0 0 0 c1t0d0 ONLINE 0 0 0 c6t1d0 ONLINE 0 0 0 c7t0d0 ONLINE 0 0 0 c8t0d0 ONLINE 0 0 0 << c5t1d0 ONLINE 0 0 0 c0t1d0 ONLINE 0 0 0 c5t2d0 ONLINE 0 0 0 c6t2d0 ONLINE 0 0 0 c7t1d0 ONLINE 0 0 0 c8t1d0 ONLINE 0 0 0 << -- Amer Ather Senior Staff Engineer Solaris Kernel Global Services Delivery amer.ather at Sun.COM 408-276-9780 (x19780) " If you fail to prepare, prepare to fail"