similar to: Issue with adding existing EFI disks to a zpool

Displaying 20 results from an estimated 1200 matches similar to: "Issue with adding existing EFI disks to a zpool"

2010 May 07
0
confused about zpool import -f and export
Hi, all, I think I''m missing a concept with import and export. I''m working on installing a Nexenta b134 system under Xen, and I have to run the installer under hvm mode, then I''m trying to get it back up under pv mode. In that process the controller names change, and that''s where I''m getting tripped up. I do a successful install, then I boot OK,
2008 Apr 08
6
lucreate error: Cannot determine the physical boot device ...
# lucreate -n B85 Analyzing system configuration. Hi, after typing # lucreate -n B85 I get the following error: No name for current boot environment. INFORMATION: The current boot environment is not named - assigning name <BE1>. Current boot environment is named <BE1>. Creating initial configuration for primary boot environment <BE1>. ERROR: Unable to determine major and
2008 Jul 07
4
getting opensolaris b85 domU to start
Folks: I am unable to get my opensolaris sn-b85 domU to start up in an opensolaris sn-b85 dom0. I am booted to xvm. Other systems, such as Fedora 8, WinXP Pro, W2k3 server, W2k8 server and Windows Vista startup normally. I have used the following to build the domU: # mkfile -n 25g /solhvm/solhvm.img # ls -l /solhvm total 269 -rw------T 1 root root 26843545600 Jul 7 10:37 solhvm.img
2010 Sep 07
3
zpool create using whole disk - do I add "p0"? E.g. c4t2d0 or c42d0p0
I have seen conflicting examples on how to create zpools using full disks. The zpool(1M) page uses "c0t0d0" but OpenSolaris Bible and others show "c0t0d0p0". E.g.: zpool create tank raidz c0t0d0 c0t1d0 c0t2d0 c0t3d0 c0t4d0 c0t5d0 zpool create tank raidz c0t0d0p0 c0t1d0p0 c0t2d0p0 c0t3d0p0 c0t4d0p0 c0t5d0p0 I have not been able to find any discussion on whether (or when) to
2006 Jul 19
1
Q: T2000: raidctl vs. zpool status
Hi all, IHACWHAC (I have a colleague who has a customer - hello, if you''re listening :-) who''s trying to build and test a scenario where he can salvage the data off the (internal ?) disks of a T2000 in case the sysboard and with it the on-board raid controller dies. If I understood correctly, he replaces the motherboard, does some magic to get the raid config back, but even
2007 Feb 03
4
Which label a ZFS/ZPOOL device has ? VTOC or EFI ?
Hi All, ZPOOL / ZFS commands writes EFI label on a device if we create ZPOOL/ZFS fs on it. Is it true ? I formatted a device with VTOC lable and I created a ZFS file system on it. Now which label the ZFS device has ? is it old VTOC or EFI ? After creating the ZFS file system on a VTOC labeled disk, I am seeing the following warning messages. Feb 3 07:47:00 scoobyb
2009 Jan 20
2
hot spare not so hot ??
I have configured a test system with a mirrored rpool and one hot spare. I powered the systems off, pulled one of the disks from rpool to simulate a hardware failure. The hot spare is not activating automatically. Is there something more i should have done to make this work ? pool: rpool state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for
2010 Jun 30
1
zfs rpool corrupt?????
Hello, Has anyone encountered the following error message, running Solaris 10 u8 in an LDom. bash-3.00# devfsadm devfsadm: write failed for /dev/.devfsadm_dev.lock: Bad exchange descriptor bash-3.00# zpool status -v rpool pool: rpool state: DEGRADED status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in
2006 Nov 07
2
xenU crashes on linux
any idea what the reason for this error is? [2006-11-07 13:59:31 xend 3167] DEBUG (__init__:1072) exception looking up device number for 0: [Errno 2] No such file or directory: ''/dev/0'' my config file looks like this: name = "solaris" memory = "400" kernel = "/solaris/platform/i86xen/kernel/amd64/unix" extra =
2008 Dec 23
1
device error without a check sum error ?
I have a system running in a VM with a root pool. The root pool occasionally shows a fairly stern warning. This warning comes with no check sum errors. bash-3.00# zpool status -vx pool: rpool state: ONLINE status: One or more devices has experienced an error resulting in data corruption. Applications may be affected. action: Restore the file in question if possible. Otherwise restore
2010 Apr 21
3
Adding a higher level partition to ZFS pool
Hi all, I would like to add a new partition to my ZFS pool but it looks like it''s more stricky than expected. The layout of my disk is the following: - first partition for Windows. I want to keep it. (no formatting !) - second partition for OpenSolaris.This is where I have all the Solaris slices (c0d0s0 etc). I have a single ZFS pool. OpenSolaris boots on ZFS. - third partition: a FAT
2006 Mar 30
39
Proposal: ZFS Hot Spare support
As mentioned last night, we''ve been reviewing a proposal for hot spare support in ZFS. Below you can find a current draft of the proposed interfaces. This has not yet been submitted for ARC review, but comments are welcome. Note that this does not include any enhanced FMA diagnosis to determine when a device is "faulted". This will come in a follow-on project, of which some
2009 Feb 11
8
Write caches on X4540
We''re using some X4540s, with OpenSolaris 2008.11. According to my testing, to optimize our systems for our specific workload, I''ve determined that we get the best performance with the write cache disabled on every disk, and with zfs:zfs_nocacheflush=1 set in /etc/system. The only issue is setting the write cache permanently, or at least quickly. Right now, as it is,
2008 Oct 26
9
vncviewer reporting connection refused (146)
I''m running snv93. I''m trying to install a windows hvm domU using the virt-install command: /usr/bin/virt-install -n windowsts -r 2048 -s 20 -f /data/domU/windowsts --hvm --vnc -c /dev/dsk/c0t0d0s2 The vncviewer is unable to open a connection to the xend VNC server and is failing with the Connection refused (146) error message. xend is set up with the following properties:
2008 Apr 02
1
delete old zpool config?
Hi experts zpool import shows some weird config of an old zpool bash-3.00# zpool import pool: data1 id: 7539031628606861598 state: FAULTED status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: http://www.sun.com/msg/ZFS-8000-3C config: data1 UNAVAIL insufficient replicas
2011 Jan 28
2
ZFS root clone problem
(for some reason I cannot find my original thread..so I''m reposting it) I am trying to move my data off of a 40gb 3.5" drive to a 40gb 2.5" drive. This is in a Netra running Solaris 10. Originally what I did was: zpool attach -f rpool c0t0d0 c0t2d0. Then I did an installboot on c0t2d0s0. Didnt work. I was not able to boot from my second drive (c0t2d0). I cannot remember
2007 Apr 19
14
Permanently removing vdevs from a pool
Is it possible to gracefully and permanently remove a vdev from a pool without data loss? The type of pool in question here is a simple pool without redundancies (i.e. JBOD). The documentation mentions for instance offlining, but without going into the end results of doing that. The thing I''m looking for is an option to evacuate, for the lack of a better word, the data from a specific
2009 Oct 13
14
How to resize ZFS partion or add a new one?
Hi, I have the following partions on my laptop, Inspiron 6000, from fdisk: 1 Other OS 0 11 12 0 2 EXT LBA 12 2561 2550 26 3 Active Solaris2 2562 9728 7167 74 First one is for Dell utilities. Second one is NTFS and the third is ZFS. I am currently using OpenSolaris 2009.06
2010 May 02
8
zpool mirror (dumb question)
Hi there! I am new to the list, and to OpenSolaris, as well as ZPS. I am creating a zpool/zfs to use on my NAS server, and basically I want some redundancy for my files/media. What I am looking to do, is get a bunch of 2TB drives, and mount them mirrored, and in a zpool so that I don''t have to worry about running out of room. (I know, pretty typical I guess). My problem is, is that
2010 Feb 12
13
SSD and ZFS
Hi all, just after sending a message to sunmanagers I realized that my question should rather have gone here. So sunmanagers please excus ethe double post: I have inherited a X4140 (8 SAS slots) and have just setup the system with Solaris 10 09. I first setup the system on a mirrored pool over the first two disks pool: rpool state: ONLINE scrub: none requested config: NAME