similar to: vdev cache - comments in the source

Displaying 20 results from an estimated 400 matches similar to: "vdev cache - comments in the source"

2007 May 29
6
Deterioration with zfs performace and recent zfs bits?
Has anyone else noticed a significant zfs performance deterioration when running recent opensolaris bits? My 32-bit / 768 MB Toshiba Tecra S1 notebook was able to do a full opensolaris release build in ~ 4 hours 45 minutes (gcc shadow compilation disabled; using an lzjb compressed zpool / zfs on a single notebook hdd p-ata drive). After upgrading to 2007-05-25 opensolaris release bits
2007 Dec 12
1
6604198 - single thread for compression
Hello zfs-discuss, http://sunsolve.sun.com/search/document.do?assetkey=1-1-6604198-1 Is there a patch for S10? I thought it''s been fixed. -- Best regards, Robert mailto:rmilkowski at task.gda.pl http://milek.blogspot.com
2008 Apr 18
1
lots of small, twisty files that all look the same
A customer has a zpool where their spectral analysis applications create a ton (millions?) of very small files that are typically 1858 bytes in length. They''re using ZFS because UFS consistently runs out of inodes. I''m assuming that ZFS aggregates these little files into recordsize (128K?) blobs for writes. This seems to go reasonably well amazingly enough. Reads are a
2008 Feb 20
12
no luck with Xen....
Perhaps someone has ideas on this topic, a recent attempt to play with Xen was a reather unlucky event, all what I will be able to demonstrate on that system is a PV ONNV domU which ikely will not be very attractive to the audience :( HW: U40M2, 2 x 2 core AMD revF procs, 8GB MEM, Phonix BIOS 1.5 (latest) 1 x 200GB internal SATA drive SW: dom0 ONNV build 82, latest VirtManager from
2010 Oct 23
2
No ACL inheritance with aclmode=passthrough in onnv-134
Hi list, while preparing for the changed ACL/mode_t mapping semantics coming with onnv-147 [1], I discovered that in onnv-134 on my system ACLs are not inherited when aclmode is set to passthrough for the filesystem. This very much puzzles me. Example: $ uname -a SunOS os 5.11 snv_134 i86pc i386 i86pc $ pwd /Volumes/ACLs/dir1 $ zfs list | grep /Volumes rpool/Volumes 7,00G 39,7G 6,84G
2008 May 14
3
tcptop_snv connp->conn_tcp != connp->conn_proto_priv.cp_tcp
Tried to start up tcptop_snv on snv87 x86 and got <i>dtrace: failed to compile script /dev/fd/11: line 168: conn_tcp is not a member of struct conn_s</i> Line 288 has this line: <i>this->tcpp = (tcp_t *)this->connp->conn_tcp;</i> which, after looking at line 168 of <a
2006 Mar 03
5
flag day: ZFS on-disk format change
Summary: If you use ZFS, do not downgrade from build 35 or later to build 34 or earlier. This putback (into Solaris Nevada build 35) introduced a backwards- compatable change to the ZFS on-disk format. Old pools will be seamlessly accessed by the new code; you do not need to do anything special. However, do *not* downgrade from build 35 or later to build 34 or earlier. If you do so, some of
2009 Nov 03
2
SunOS neptune 5.11 snv_127 sun4u sparc SUNW, Sun-Fire-880
I just went through a BFU update to snv_127 on a V880 : neptune console login: root Password: Nov 3 08:19:12 neptune login: ROOT LOGIN /dev/console Last login: Mon Nov 2 16:40:36 on console Sun Microsystems Inc. SunOS 5.11 snv_127 Nov. 02, 2009 SunOS Internal Development: root 2009-Nov-02 [onnv_127-tonic] bfu''ed from /build/archives-nightly-osol/sparc on 2009-11-03 I have [
2008 Apr 08
1
Re: [ldoms-discuss] run latest build on b84
[ Redirecting to xen-discuss@opensolaris.org ] xiao liu wrote: > I installed Nevada b84 on my laptop then built latest onnv-gate > on it. > > After I installed it using Install and try to boot, > it panic for don''t know "/xpvd" (hvmboot_rootconf failed) > > Seems b84 don''t support xpvd/xpv, What should I do now, > should I use BFU instead?
2008 Dec 26
19
separate home "partition"?
(i use the term loosely because i know that zfs likes whole volumes better) when installing ubuntu, i got in the habit of using a separate partition for my home directory so that my data and gnome settings would all remain intact when i reinstalled or upgraded. i''m running osol 2008.11 on an ultra 20, which has only two drives. i''ve got all my data located in my home directory,
2006 Nov 16
2
Porting ZFS, trouble with nvpair
Hi. I thought I''d take a stab at the first steps of porting ZFS to Darwin. I realize there are rumors that Apple is already doing this, but my contact at Apple has yet to get back to me to verify this. In the meantime, I wanted to see how hard it would be. I started with libzfs, and promptly ran into issues with libnvpair. It wants sys/nvpair.h, but I can''t find that in the
2010 Mar 02
2
dedup source code
Hello ZFS experts: I would like to study ZFS de-duplication feature. Can someone please let me know which directory/files I should be looking at? Thanks in advance. -- This message posted from opensolaris.org
2007 Jul 26
4
Does iSCSI target support SCSI-3 PGR reservation ?
Does opensolaris iSCSI target support SCSI-3 PGR reservation ? My goal is to use the iSCSI LUN created by [1] or [2] as a quorum device for a 3-node suncluster. [1] zfs set shareiscsi=on <storage-pool/zfs volume name> [2] iscsitadm create target ..... Thanks, -- leon This message posted from opensolaris.org
2009 Nov 02
24
dedupe is in
Deduplication was committed last night by Mr. Bonwick: > Log message: > PSARC 2009/571 ZFS Deduplication Properties > 6677093 zfs should have dedup capability http://mail.opensolaris.org/pipermail/onnv-notify/2009-November/010683.html Via c0t0d0s0.org.
2007 Nov 21
0
Using DTrace to analyse ON build
This is from a discussion on opensolaris-tools that is somewhat relevant for DTrace community. I wrote a simple D script (http://www.opensolaris.org/os/project/onnv/onnv_build/faster_builds/tools/dtrace/dirtrace.d) that collects data for each dmake process during the Solaris build. After that the data is post-processed by a Perl script
2010 Jun 11
9
Are recursive snapshot destroy and rename atomic too?
In another thread recursive snapshot creation was found atomic so that it is done quickly, and more important, all at once or nothing at all. Do you know if recursive destroying and renaming of snapshots are atomic too? Regards Henrik Heino
2009 Feb 18
4
tracing aio syscalls
Hi all, Is there some documentation or some example on how to interpret the arg0 .. arg<n> for the aioread, aiowrite, aiowait syscalls? The system call name for all three seems to be "kaio". Michael === Michael Mueller ================== Tel. + 49 8171 63600 Fax. + 49 8171 63615 Web: http://www.michael-mueller-it.de ======================================
2007 Oct 09
4
dom0 boot panic after bfu to b75
After BFU-ing my system to b75 I ended up with a panicing system when booting into Xen: grub> #---------- ADDED BY BOOTADM - DO NOT EDIT ---------- grub> title Solaris on Xen grub> kernel$ /boot/$ISADIR/xen.gz grub> module$ /platform/i86xpv/kernel/$ISADIR/unix /platform/i86xpv/kernel/$ISADIR/unix -B console=ttyb grub> module$ /platform/i86pc/$ISADIR/boot_archive grub>
2009 Mar 08
2
[Bug 866] ssh(1) is too picky about unknown options in ~/.ssh/config
https://bugzilla.mindrot.org/show_bug.cgi?id=866 --- Comment #17 from Olav Morken <olavmrk at gmail.com> 2009-03-09 06:21:16 --- Created an attachment (id=1610) --> (http://bugzilla.mindrot.org/attachment.cgi?id=1610) Patch which allows OpenSSH to ignore unknown options. This is a patch which implements alternative 1 from Josh Triplett. This patch makes ssh ignore all unknown
2010 Aug 15
2
Is the error threshold for a degraded device configurable?
I look after an x4500 for a client and wee keep getting drives marked as degraded with just over 20 checksum errors. Most of these errors appear to be driver or hardware related and thier frequency increases during a resilver, which can lead to a death spiral. The increase in errors within a vdev during a resilver (I recently had three drives in an 8 drive raidz vdev "degraded")