similar to: Extremely poor ZFS perf and other observations

Displaying 20 results from an estimated 1000 matches similar to: "Extremely poor ZFS perf and other observations"

2008 Apr 18
2
plockstat: failed to add to aggregate: Abort due to drop
when check java process lock statistics, plockstat failed, please see below: # prstat -mLp 21162 PID USERNAME USR SYS TRP TFL DFL LCK SLP LAT VCX ICX SCL SIG PROCESS/LWPID 21162 7677 0.9 0.1 0.0 0.0 0.0 99 0.0 0.3 83 89 215 0 java/81 21162 7677 0.3 0.1 0.0 0.0 0.0 0.0 99 0.2 106 33 305 0 java/35 21162 7677 0.1 0.0 0.0 0.0 0.0 100 0.0 0.1 79 6 85 0 java/59
2012 Aug 08
1
How to build vfs_zfsacl module on Linux platform?
Solaris OS has ZFS as primary filesystem, that has own implementation of ACL (NFS v4 ACL). Samba uses zfsacl and acl_xattr modules for converting NT ACL to ZFS ACL and extended attributes of file in this OS. ZfsOnLinux team ported ZFS as a number of kernel-linked modules to Linux OS, so ZFS on Linux supports NFS ACL, deduplication, compresson and other features of Solaris ZFS v.28. But NFS ACL is
2010 Jun 17
2
feature request: lzma compresion (7zip)
Hello, many linux SW is starting to implement new lzma compresson instrad of old zlib (gzip) od bzip2. lzma is default comrpession in very good compression SW 7-zip, which is faster and have higher compression ratio then bzip2 or rar. Currently its probalby the best compressor in therms of compression and decompression speed / compression ratio. In linux there is GNU lzma SW which implements
2007 Sep 23
3
ext3 file system becoming read only
Hi In our office environment few servers mostly database servers and yesterday it happened for one application server(first time) the partion is getting "read only". I was checking the archives, found may be similar kind of issues in the 2007-July archives. But how it has been solved if someone describes me that will be really helpful. In our case, just at the problem started found
2004 Nov 29
0
Re: Mode context extremely poor performance and varyio
Stephen, all, Thank u very much for your answer and i wish you an happy thanksgiving. Currently we tried to migrate our base on RAW devices already using e41smp, SecurePath (HP) 3C and Qlogic 7.00. The average response time of our sql request on ia32 is 0.32Sec very regularly. On our Itanium on OCFS that's very randomly between 1 and 15Sec, and on RAW between <1s and 3 sec. David is an
2004 Nov 29
0
Re: Mode context extremely poor performance and varyio
Stephen, all, Thank u very much for your answer and i wish you an happy thanksgiving. Currently we tried to migrate our base on RAW devices already using e41smp, SecurePath (HP) 3C and Qlogic 7.00. The average response time of our sql request on ia32 is 0.32Sec very regularly. On our Itanium on OCFS that's very randomly between 1 and 15Sec, and on RAW between <1s and 3 sec. David is an
2007 Jan 08
1
Extremely poor rsync performance on very large files (near 100GB and larger)
I've been playing with rsync and very large files approaching and surpassing 100GB, and have found that rsync has excessively very poor performance on these very large files, and the performance appears to degrade the larger the file gets. The problem only appears to happen when the file is being "updated", that is, when it already exists on the receiving side. For instance,
2004 Nov 22
3
Mode context extremely poor performance.
Hi all, I curently have a big problem. One request (listed above) using context take up to 1000 more time than the on RAW or ext2 database. I have ran this request on a single IA32 machine with Redhat and dbf on ext2. The average reponse time is less than a sec. The same request on RAC 4 nodes cluster on RAW take the same average time. On ext2 idem. But on OCFS it took up to 15 sec randomly
2004 Nov 22
3
Mode context extremely poor performance.
Hi all, I curently have a big problem. One request (listed above) using context take up to 1000 more time than the on RAW or ext2 database. I have ran this request on a single IA32 machine with Redhat and dbf on ext2. The average reponse time is less than a sec. The same request on RAC 4 nodes cluster on RAW take the same average time. On ext2 idem. But on OCFS it took up to 15 sec randomly
2008 Jan 10
2
FreeBSD tar errors on valid empty tar.gz
Seems our current libarchive? That support FreeBSD's tar implementation has a bug where it can create archives it cant read back. This can be seen by simply creating an empty tar.gz file and then trying to expand or list it. In doing the above you get the following error: tar: Unrecognized archive format: Inappropriate file type or format N.B. gtar can list and expand the created file
2012 Feb 06
6
schily tools
Hey folks, I'm reading up on gtar for tape archiving and it sounds kind of nasty and not something I really want to rely on. It looks like star from the schily tools is preferred. I'm using Centos (and RHEL) 5.7 which seems to have star but not sdd. Which leads me to believe that the Schily tools are maybe a bit "rogue" My basic requirement with what I'm doing is to use
2007 Feb 14
5
Need help making lsof work with ZFS
I contacted the author of ''lsof'' regarding the missing ZFS support. The command works but fails to display any files that are opened by the process in a ZFS filesystem. He indicates that the required ZFS kernel structure definitions (header files) are not shipped with the OS. He further indicated that he rummaged through the OpenSolaris source tree and the files doesn''t
2006 Oct 16
11
Configuring a 3510 for ZFS
Hi folks, Myself and a colleague are currently involved in a prototyping exercise to evaluate ZFS against our current filesystem. We are looking at the best way to arrange the disks in a 3510 storage array. We have been testing with the 12 disks on the 3510 exported as "nraid" logical devices. We then configured a single ZFS pool on top of this, using two raid-z arrays. We are getting
2002 Feb 04
1
installing version 1.4.1
Hello people, me again ... Just an FYI this time, I think: 0[209]1 sunray1:/contrib > gtar zxf R-1.4.1.tgz tar: R-1.4.1/date-stamp: Could not create file: Permission denied gtar: Error exit delayed from previous errors 2[210]1 sunray1:/contrib > gtar zxvf R-1.4.1.tgz "R-1.4.1/date-stamp" R-1.4.1/date-stamp gtar: R-1.4.1/date-stamp: Could not create file: Permission denied gtar:
2010 Dec 15
3
having trouble with puppet 0.25.5 on openbsd 4.8 on amd64
My manifest includes this snippet class app_client_openbsd { $app_client_pkgs_obsd = [ ''glib2-2.24.1p2'', ''gtar-1.23p1'' ] package { "${app_client_pkgs_obsd}": ensure => ''installed'', source => "http://${installserver}/openbsd/${operatingsystemrelease}/packages/${hardwaremodel}/${name}.tgz", } This
2007 Feb 27
16
understanding zfs/thunoer "bottlenecks"?
Currently I''m trying to figure out the best zfs layout for a thumper wrt. to read AND write performance. I did some simple mkfile 512G tests and found out, that per average ~ 500 MB/s seems to be the maximum on can reach (tried initial default setup, all 46 HDDs as R0, etc.). According to http://www.amd.com/us-en/assets/content_type/DownloadableAssets/ArchitectureWP_062806.pdf I would
2010 Dec 09
1
Extremely poor write performance, but read appears to be okay
Hello, I'm writing from the otherside of the world from where my systems are, so details are coming in slow. We have a 6TB OCFS2 volume across 20 or so nodes all running OEL5.4 running ocfs2-1.4.4. The system has worked fairly well for the last 6-8 months. Something has happened over the last few weeks which has driven write performance nearly to a halt. I'm not sure how to proceed, and
2006 Mar 29
3
ON 20060327 and upcoming solaris 10 U2 / coreutils
So, I noticed that a lot of the fixes discussed here recently, including the ZFS/NFS interaction bug fixes and the deadlock fix has made it into 20060327 that was released this morning. My question is whether we''ll see all these up to the minute bug fixes in the Solaris 10 update that brings ZFS to that product, or if there is a specific date where no further updates will make it in to
2006 Mar 28
43
zfs and backup applications
Hi, I was wondering if there have been any conversations with backup vendors like Veritas or EMC regarding better integration with ZFS. While I understand they can use the "native" mode of reading files from the filesystem, it would be great if there were agents that had options like making a snapshot and storing a "zfs backup" datastream that could be used for zfs restore.
2006 Nov 28
1
GNU Tar vulnerability
Please, note: http://secunia.com/advisories/23115/ A port maintainer CC'ed. -- Dixi. Sem.