search for: richardel

Displaying 20 results from an estimated 79 matches for "richardel".

Did you mean: richardet
2010 Apr 28
6
Compellant announces zNAS
Today, Compellant announced their zNAS addition to their unified storage line. zNAS uses ZFS behind the scenes. http://www.compellent.com/Community/Blog/Posts/2010/4/Compellent-zNAS.aspx Congrats Compellant! -- richard ZFS storage and performance consulting at http://www.RichardElling.com ZFS training on deduplication, NexentaStor, and NAS performance Las Vegas, April 29-30, 2010 http://nexenta-vegas.eventbrite.com
2011 Oct 26
1
Problem running zilstat script on a core install server.
We cannot run the zilstat utility (from http://www.richardelling.com/Home/scripts-and-programs-1/zilstat) on an Solaris 10 u9 / x86 installed according SUNWCreq (which is a core install) while it works fine on a similar config but installed with all packages (SUNWCXall) So we know it works in this environment but it seems we miss an essential package here....
2010 Mar 29
2
pool won''t import
root at cs6:~# zpool import pool: content3 id: 14184872052409584084 state: FAULTED status: The pool metadata is corrupted. action: The pool cannot be imported due to damaged devices or data. The pool may be active on another system, but can be imported using the ''-f'' flag. see: http://www.sun.com/msg/ZFS-8000-72 config: content3
2010 May 04
8
iscsitgtd failed request to share on zpool import after upgrade from b104 to b134
Hi, I am posting my question to both storage-discuss and zfs-discuss as I am not quite sure what is causing the messages I am receiving. I have recently migrated my zfs volume from b104 to b134 and upgraded it from zfs version 14 to 22. It consist of two zvol''s ''vol01/zvol01'' and ''vol01/zvol02''. During zpool import I am getting a non-zero exit code,
2011 Dec 02
14
LSI 3GB HBA SAS Errors (and other misc)
During the diagnostics of my SAN failure last week we thought we had seen a backplane failure due to high error counts with ''lsiutil''. However, even with a new backplane and ruling out failed cards (MPXIO or singular) or bad cables I''m still seeing my error count with LSIUTIL increment. I''ve got no disks attached to the array right now so I''ve also
2012 Apr 06
6
Seagate Constellation vs. Hitachi Ultrastar
Happy Friday, List! I''m spec''ing out a Thumper-esque solution and having trouble finding my favorite Hitachi Ultrastar 2TB drives at a reasonable post-flood price. The Seagate Constellations seem pretty reasonable given the market circumstances but I don''t have any experience with them. Anybody using these in their ZFS systems and have you had good luck? Also, if
2010 Feb 15
3
zfs questions wrt unused blocks
Gents, We want to understand the mechanism of zfs a bit better. Q: what is the design/algorithm of zfs in terms of reclaiming unused blocks? Q: what criteria is there for zfs to start reclaiming blocks Issue at hand is an LDOM or zone running in a virtual (thin-provisioned) disk on a NFS server and a zpool inside that vdisk. This vdisk tends to grow in size even if the user writes and deletes
2010 Mar 06
3
Monitoring my disk activity
Recently, I''m benchmarking all kinds of stuff on my systems. And one question I can''t intelligently answer is what blocksize I should use in these tests. I assume there is something which monitors present disk activity, that I could run on my production servers, to give me some statistics of the block sizes that the users are actually performing on the production server.
2012 Jul 02
14
HP Proliant DL360 G7
Hello, Has anyone out there been able to qualify the Proliant DL360 G7 for your Solaris/OI/Nexenta environments? Any pros/cons/gotchas (vs. previous generation HP servers) would be greatly appreciated. Thanks in advance! -Anh
2012 Dec 12
20
Solaris 11 System Reboots Continuously Because of a ZFS-Related Panic (7191375)
I''ve hit this bug on four of my Solaris 11 servers. Looking for anyone else who has seen it, as well as comments/speculation on cause. This bug is pretty bad. If you are lucky you can import the pool read-only and migrate it elsewhere. I''ve also tried setting zfs:zfs_recover=1,aok=1 with varying results. http://docs.oracle.com/cd/E26502_01/html/E28978/gmkgj.html#scrolltoc
2009 Dec 04
30
ZFS send | verify | receive
If there were a ?zfs send? datastream saved someplace, is there a way to verify the integrity of that datastream without doing a ?zfs receive? and occupying all that disk space? I am aware that ?zfs send? is not a backup solution, due to vulnerability of even a single bit error, and lack of granularity, and other reasons. However ... There is an attraction to ?zfs send? as an augmentation to the
2010 Mar 05
17
why L2ARC device is used to store files ?
Greeting All I have create a pool that consists oh a hard disk and a ssd as a cache zpool create hdd c11t0d0p3 zpool add hdd cache c8t0d0p0 - cache device I ran an OLTP bench mark to emulate a DMBS One I ran the benchmark, the pool started create the database file on the ssd cache device ??????????? can any one explain why this happening ? is not L2ARC is used to absorb the evicted data
2010 Sep 29
10
Resliver making the system unresponsive
This must be resliver day :) I just had a drive failure. The hot spare kicked in, and access to the pool over NFS was effectively zero for about 45 minutes. Currently the pool is still reslivering, but for some reason I can access the file system now. Resliver speed has been beaten to death I know, but is there a way to avoid this? For example, is more enterprisy hardware less susceptible to
2010 May 08
6
Mirrored Servers
Lets say I have two servers, both running opensolaris with ZFS. I basically want to be able to create a filesystem where the two servers have a common volume, that is mirrored between the two. Meaning, each server keeps an identical, real time backup of the other''s data directory. Set them both up as file servers, and load balance between the two for incoming requests. How would anyone
2010 Mar 01
1
ARC & Maxphys & recordsize
Greeting ALL Can someone explain to me why I was successfully able issue (through application) an I/Os of 128K each (monitored by DTrace) while my Maxphys=56K only ?????? My understanding is that Maxphys is the max I/O size that the storage device can handle for a single I/O, it has been well documented that if and application issued an I/O larger than Maxphys, it would be break down to
2010 Mar 11
1
zpool iostat / how to tell if your iop bound
What is the best way to tell if your bound by the number of individual operations per second / random io? "zpool iostat" has an "operations" column but this doesn''t really tell me if my disks are saturated. Traditional "iostat" doesn''t seem to be the greatest place to look when utilizing zfs. Thanks, Chris -------------- next part -------------- An
2010 Mar 12
2
ZFS error while enabling DIRECT_IO for the DB chunks
Hi, We are using Solaris 10 update 7 with ZFS file system.And using the machine for informix db. Solaris Patch level Generic_142900-02 (Dec 09 PatchCluster release) Informix DB version 11.5FC6 We are facing an issue while enabling DIRECT_IO for the DB chunks. The error message which appears in the online.log file is "Direct I/O cannot be used for the chunk file <file_name>"
2010 Mar 02
2
dedup source code
Hello ZFS experts: I would like to study ZFS de-duplication feature. Can someone please let me know which directory/files I should be looking at? Thanks in advance. -- This message posted from opensolaris.org
2012 Sep 27
0
reminder: ZFS day next Tuesday
...f you can''t make it, be sure to register for the streaming video feeds. Details at: www.zfsday.com Be sure to prep your ZFS war stories for the beer bash afterwards -- thanks Delphix! -- richard -- illumos Day & ZFS Day, Oct 1-2, 2012 San Fransisco www.zfsday.com Richard.Elling at RichardElling.com +1-760-896-4422 -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20120927/caa8ff00/attachment.html>
2010 Mar 02
3
compressed root pool at installation time with flash archive predeployment script
I was trying to think of a way to set compression=on at the beginning of a jumpstart. The only idea I''ve come up with is to do so with a flash archive predeployment script. Has anyone else tried this approach? Thanks, Chad -------------- next part -------------- An HTML attachment was scrubbed... URL: