similar to: ZFS with SAN Disks and mutipathing

Displaying 20 results from an estimated 500 matches similar to: "ZFS with SAN Disks and mutipathing"

2008 Feb 08
4
List of supported multipath drivers
Where can I find a list of supported multipath drivers for ZFS? Keith McAndrew Senior Systems Engineer Northern California SUN Microsystems - Data Management Group <mailto:Keith.McAndrew at SUN.com> Keith.McAndrew at SUN.com 916 715 8352 Cell CONFIDENTIALITY NOTICE The information contained in this transmission may contain privileged and confidential information of SUN
2007 Jul 13
28
ZFS and powerpath
How much fun can you have with a simple thing like powerpath? Here''s the story: I have a (remote) system with access to a couple of EMC LUNs. Originally, I set it up with mpxio and created a simple zpool containing the two LUNs. It''s now been reconfigured to use powerpath instead of mpxio. My problem is that I can''t import the pool. I get: pool: ###### id:
2006 Oct 18
5
ZFS and IBM sdd (vpath)
Hello, I am trying to configure ZFS with IBM sdd. IBM sdd is like powerpath, MPXIO or VxDMP. Here is the error message when I try to create my pool: bash-3.00# zpool create tank /dev/dsk/vpath1a warning: device in use checking failed: No such device internal error: unexpected error 22 at line 446 of ../common/libzfs_pool.c bash-3.00# zpool create tank /dev/dsk/vpath1c cannot open
2006 Dec 21
1
Dtrace + mpxio
Hi ! I am using MPXIO for doing multipathing across four HBA''s. Now I need to know how much IO/sec each HBA card is serving. With EMC powerpath installed, we could get this using powermt display every=5. is there any way I can use Dtrace to extract this output ? I guess MPXIO just shows one Device name to the OS so dev__pathname won''t be able to tell me per HBA i/O. Regds,
2008 Mar 20
7
ZFS panics solaris while switching a volume to read-only
Hi, I just found out that ZFS triggers a kernel-panic while switching a mounted volume into read-only mode: The system is attached to a Symmetrix, all zfs-io goes through Powerpath: I ran some io-intensive stuff on /tank/foo and switched the device into read-only mode at the same time (symrdf -g bar failover -establish). ZFS went ''bam'' and triggered a Panic: WARNING: /pci at
2007 Apr 23
14
concatination & stripe - zfs?
I want to configure my zfs like this : concatination_stripe_pool : concatination lun0_controller0 lun1_controller0 concatination lun2_controller1 lun3_controller1 1. There is any option to implement it in ZFS? 2. there is other why to get the same configuration? thanks This message posted from opensolaris.org
2007 Nov 13
4
sd_max_throttle
Hello, we are using hardware array and its vendor recommends the following setting in /etc/system: set sd:sd_max_throttle = <value> or set ssd:ssd_max_throttle = <value> Is it possible to monitor *somehow* whether the variable becomes sort of bottleneck ? Or how its value influences io traffic ? Regards przemol
2009 Jan 13
12
OpenSolaris better Than Solaris10u6 with requards to ARECA Raid Card
Under Solaris 10 u6 , No matter how I configured my ARECA 1261ML Raid card I got errors on all drives that result from SCSI timeout errors. yoda:~ # tail -f /var/adm/messages Jan 9 11:03:47 yoda.asc.edu scsi: [ID 107833 kern.notice] Requested Block: 239683776 Error Block: 239683776 Jan 9 11:03:47 yoda.asc.edu scsi: [ID 107833 kern.notice] Vendor: Seagate
2007 Jun 15
3
zfs and EMC
Hi there, have a strange behavior if i?ll create a zfs pool at an EMC PowerPath pseudo device. I can create a pool on emcpower0a but not on emcpower2a zpool core dumps with invalid argument .... ???? Thats my second maschine with powerpath and zfs the first one works fine, even zfs/powerpath and failover ... Is there anybody who has the same failure and a solution ? :) Greets Dominik
2005 Aug 04
3
Ocfs and EMC Powerpath
A couple years ago, we moved to Oracle RAC on Linux using ocfs that is SAN attached to an EMC CLARiiON. At the time, there was a reason that we did NOT use EMC's Powerpath (I just can't recall what that reason was). What I'd like to know is if there are any issues with introducing Powerpath now. * RHEL 2.1 AS * 2.4.9-e.38enterprise * ocfs-2.4.9-e-enterprise-1.0.13-1 * EMC
2010 Aug 09
2
ZFS with EMC PowerPath
On some machines running PowerPath, there''s sometimes issues after an update/upgrade of the PowerPath software. Sometimes the pseudo devices get remapped and change names. ZFS appears to handle it OK, however sometimes it then references half native device names and half the emcpower pseudo device names. So the SA can easily be confused. Is there a way to tell ZFS which device names
2009 Jan 09
7
Desperate question about MPXIO with ZFS-iSCSI
I''m trying to set up a iscsi connection (with MPXIO) between my Vista64 workstation and a ZFS storage machine running OpenSolaris 10 (forget the exact version). On the ZFS machines, I have two NICS. NIC #1 is 192.168.1.102, and NIC #2 is 192.168.2.102. The NICs are connected to two separate switches serving two separate IP spaces. On my Vista64 machine, I also have two NICs connected in
2010 Nov 05
3
ZFS vs mpxio vs cfgadm in Solaris.
Folks, I''m trying to figure out whether we should give ZFS / mpxio a shot on one of our research servers, or simply skip it (as we have previously). In Nov 2009 Cindy responded to a thread concerning ZFS device issues, cfgadm, and mpxio: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-November/033496.html I''ve got an x2270 with the Sun EZ-SAS HBA and external SATA
2006 Dec 22
6
Re: Difference between ZFS and UFS with one LUN froma SAN
This may not be the answer you''re looking for, but I don''t know if it''s something you''ve thought of. If you''re pulling a LUN from an expensive array, with multiple HBA''s in the system, why not run mpxio? If you ARE running mpxio, there shouldn''t be an issue with a path dropping. I have the setup above in my test lab and pull cables
2007 Apr 19
14
Experience with Promise Tech. arrays/jbod''s?
Greetings, In looking for inexpensive JBOD and/or RAID solutions to use with ZFS, I''ve run across the recent "VTrak" SAS/SATA systems from Promise Technologies, specifically their E-class and J-class series: E310f FC-connected RAID: http://www.promise.com/product/product_detail_eng.asp?product_id=175 E310s SAS-connected RAID:
2008 Jul 09
8
Using zfs boot with MPxIO on T2000
Here is what I have configured: T2000 with OBP 4.28.6 2008/05/23 12:07 with 2 - 72 GB disks as the root disks OpenSolaris Nevada Build 91 Solaris Express Community Edition snv_91 SPARC Copyright 2008 Sun Microsystems, Inc. All Rights Reserved. Use is subject to license terms. Assembled 03 June 2008
2013 Jan 07
5
mpt_sas multipath problem?
Greetings, We''re trying out a new JBOD here. Multipath (mpxio) is not working, and we could use some feedback and/or troubleshooting advice. The OS is oi151a7, running on an existing server with a 54TB pool of internal drives. I believe the server hardware is not relevant to the JBOD issue, although the internal drives do appear to the OS with multipath device names (despite the fact
2009 Oct 27
4
EMC CX4 Clariion
Hi, We had received a new EMC storage (Clariion CX4) and the EMC analist has told us that CentOS aren't on their support list, but RHEL are :) Well anybody on the list has a CentOS host talking (iSCSI) with an EMC storage, preferably using the EMC Powerpath software, and can talk about you experience? Or I will need to buy some REHL licenses ? I'll put my hands on it only next
2007 Jul 18
1
Converting exisitng ZFS pool to MPxIO
We have a Sun v890, and I''m interested converting existing ZFS zpool from c#t#d# to MPxIO. % zpool status pool: data state: ONLINE status: ONLINE scrub: scrub completed with 0 errors on Sun Jul 15 10:58:33 2007 config: NAME STATE READ WRITE CKSUM data ONLINE 0 0 0 mirror ONLINE 0 0 0 c1t2d0
2005 Aug 31
8
problem with OCFS label
I used this command to create volume label on OCFS: mkfs.ocfs -F -b 128 -L data13 -m /oradata/data13 -u oracle -g dba -p 0775 /dev/emcpowerp1 emcpowerp is composed of /dev/sdad and /dev/sdk. It seems the above command created the same labels for /dev/emcpowerp1, /dev/sdad1 and /dev/sdk1. But when I tried to mount this ocfs filesystem by label, it gave me the following error. # mount -L data13