similar to: Compellant announces zNAS

Displaying 20 results from an estimated 500 matches similar to: "Compellant announces zNAS"

2009 Jun 15
33
compression at zfs filesystem creation
Hi, I just installed 2009.06 and found that compression isn''t enabled by default when filesystems are created. Does is make sense to have an RFE open for this? (I''ll open one tonight if need be.) We keep telling people to turn on compression. Are there any situations where turning on compression doesn''t make sense, like rpool/swap? what about rpool/dump? Thanks, ~~sa
2011 Oct 26
1
Problem running zilstat script on a core install server.
We cannot run the zilstat utility (from http://www.richardelling.com/Home/scripts-and-programs-1/zilstat) on an Solaris 10 u9 / x86 installed according SUNWCreq (which is a core install) while it works fine on a similar config but installed with all packages (SUNWCXall) So we know it works in this environment but it seems we miss an essential package here. That causes the dtrace: failed to
2012 Jul 02
14
HP Proliant DL360 G7
Hello, Has anyone out there been able to qualify the Proliant DL360 G7 for your Solaris/OI/Nexenta environments? Any pros/cons/gotchas (vs. previous generation HP servers) would be greatly appreciated. Thanks in advance! -Anh
2010 Mar 29
2
pool won''t import
root at cs6:~# zpool import pool: content3 id: 14184872052409584084 state: FAULTED status: The pool metadata is corrupted. action: The pool cannot be imported due to damaged devices or data. The pool may be active on another system, but can be imported using the ''-f'' flag. see: http://www.sun.com/msg/ZFS-8000-72 config: content3
2010 May 04
8
iscsitgtd failed request to share on zpool import after upgrade from b104 to b134
Hi, I am posting my question to both storage-discuss and zfs-discuss as I am not quite sure what is causing the messages I am receiving. I have recently migrated my zfs volume from b104 to b134 and upgraded it from zfs version 14 to 22. It consist of two zvol''s ''vol01/zvol01'' and ''vol01/zvol02''. During zpool import I am getting a non-zero exit code,
2012 Sep 27
0
reminder: ZFS day next Tuesday
If you''ve been hiding under a rock, not checking your email, then you might not have heard about the Next Big Whopper Event for ZFS Fans: ZFS Day! The agenda is now set and the teams are preparing to descend towards San Francisco''s Moscone Center vortex for a full day of ZFS. I''d love to see y''all there in person, but if you can''t make it, be sure to
2010 Feb 15
3
zfs questions wrt unused blocks
Gents, We want to understand the mechanism of zfs a bit better. Q: what is the design/algorithm of zfs in terms of reclaiming unused blocks? Q: what criteria is there for zfs to start reclaiming blocks Issue at hand is an LDOM or zone running in a virtual (thin-provisioned) disk on a NFS server and a zpool inside that vdisk. This vdisk tends to grow in size even if the user writes and deletes
2010 Mar 06
3
Monitoring my disk activity
Recently, I''m benchmarking all kinds of stuff on my systems. And one question I can''t intelligently answer is what blocksize I should use in these tests. I assume there is something which monitors present disk activity, that I could run on my production servers, to give me some statistics of the block sizes that the users are actually performing on the production server.
2010 Sep 29
10
Resliver making the system unresponsive
This must be resliver day :) I just had a drive failure. The hot spare kicked in, and access to the pool over NFS was effectively zero for about 45 minutes. Currently the pool is still reslivering, but for some reason I can access the file system now. Resliver speed has been beaten to death I know, but is there a way to avoid this? For example, is more enterprisy hardware less susceptible to
2010 Mar 12
2
ZFS error while enabling DIRECT_IO for the DB chunks
Hi, We are using Solaris 10 update 7 with ZFS file system.And using the machine for informix db. Solaris Patch level Generic_142900-02 (Dec 09 PatchCluster release) Informix DB version 11.5FC6 We are facing an issue while enabling DIRECT_IO for the DB chunks. The error message which appears in the online.log file is "Direct I/O cannot be used for the chunk file <file_name>"
2010 Mar 02
3
compressed root pool at installation time with flash archive predeployment script
I was trying to think of a way to set compression=on at the beginning of a jumpstart. The only idea I''ve come up with is to do so with a flash archive predeployment script. Has anyone else tried this approach? Thanks, Chad -------------- next part -------------- An HTML attachment was scrubbed... URL:
2010 Mar 05
17
why L2ARC device is used to store files ?
Greeting All I have create a pool that consists oh a hard disk and a ssd as a cache zpool create hdd c11t0d0p3 zpool add hdd cache c8t0d0p0 - cache device I ran an OLTP bench mark to emulate a DMBS One I ran the benchmark, the pool started create the database file on the ssd cache device ??????????? can any one explain why this happening ? is not L2ARC is used to absorb the evicted data
2011 Dec 02
14
LSI 3GB HBA SAS Errors (and other misc)
During the diagnostics of my SAN failure last week we thought we had seen a backplane failure due to high error counts with ''lsiutil''. However, even with a new backplane and ruling out failed cards (MPXIO or singular) or bad cables I''m still seeing my error count with LSIUTIL increment. I''ve got no disks attached to the array right now so I''ve also
2010 Sep 14
9
dedicated ZIL/L2ARC
We are looking into the possibility of adding a dedicated ZIL and/or L2ARC devices to our pool. We are looking into getting 4 ? 32GB Intel X25-E SSD drives. Would this be a good solution to slow write speeds? We are currently sharing out different slices of the pool to windows servers using comstar and fibrechannel. We are currently getting around 300MB/sec performance with 70-100% disk busy.
2010 Apr 05
14
Are there (non-Sun/Oracle) vendors selling OpenSolaris/ZFS based NAS Hardware?
I''ve seen the Nexenta and EON webpages, but I''m not looking to build my own. Is there anything out there I can just buy? -Kyle
2008 Jun 25
6
dm-multipath use
Are folks in the Centos community succesfully using device-mapper-multipath? I am looking to deploy it for error handling on our iSCSI setup but there seems to be little traffic about this package on the Centos forums, as far as I can tell, and there seems to be a number of small issues based on my reading the dm-multipath developer lists and related resources. -geoff Geoff Galitz Blankenheim
2010 Mar 01
1
ARC & Maxphys & recordsize
Greeting ALL Can someone explain to me why I was successfully able issue (through application) an I/Os of 128K each (monitored by DTrace) while my Maxphys=56K only ?????? My understanding is that Maxphys is the max I/O size that the storage device can handle for a single I/O, it has been well documented that if and application issued an I/O larger than Maxphys, it would be break down to
2010 Mar 11
1
zpool iostat / how to tell if your iop bound
What is the best way to tell if your bound by the number of individual operations per second / random io? "zpool iostat" has an "operations" column but this doesn''t really tell me if my disks are saturated. Traditional "iostat" doesn''t seem to be the greatest place to look when utilizing zfs. Thanks, Chris -------------- next part -------------- An
2010 Mar 02
2
dedup source code
Hello ZFS experts: I would like to study ZFS de-duplication feature. Can someone please let me know which directory/files I should be looking at? Thanks in advance. -- This message posted from opensolaris.org
2010 Apr 23
5
Data movement across filesystems within a pool
I would have thought that the file movement from one FS to another within the same pool would be almost instantaneous. Why does it take to platter for such a movement? # time cp /tmp/blockfile /pcshare/1gb-tempfile real 0m5.758s # time mv /pcshare/1gb-tempfile . real 0m4.501s Both FSs are with compression=off. /tmp is RAM. -devsk -- This message posted from opensolaris.org