similar to: Targeted build version

Displaying 20 results from an estimated 9000 matches similar to: "Targeted build version"

2010 Jan 02
27
Pool import with failed ZIL device now possible ?
Hello list, someone (actually neil perrin (CC)) mentioned in this thread: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-December/034340.html that is should be possible to import a pool with failed log devices (with or without data loss ?). >/ />/ Has the following error no consequences? />/ />/ Bug ID 6538021 />/ Synopsis Need a way to force pool startup when
2010 Jan 10
5
Repeating scrub does random fixes
I''ve been using a 5-disk raidZ for years on SXCE machine which I converted to OSOL. The only time I ever had zfs problems in SXCE was with snv_120, which was fixed. So, now I''m at OSOL snv_111b and I''m finding that scrub repairs errors on random disks. If I repeat the scrub, it will fix errors on other disks. Occasionally it runs cleanly. That it doesn''t
2010 Aug 12
6
one ZIL SLOG per zpool?
I have three zpools on a server and want to add a mirrored pair of ssd''s for the ZIL. Can the same pair of SSDs be used for the ZIL of all three zpools or is it one ZIL SLOG device per zpool? -- This message posted from opensolaris.org
2009 Nov 20
1
Using local disk for cache on an iSCSI zvol...
I''m just wondering if anyone has tried this, and what the performance has been like. Scenario: I''ve got a bunch of v20z machines, with 2 disks. One has the OS on it, and the other is free. As these are disposable client machines, I''m not going to mirror the OS disk. I have a disk server with a striped mirror zpool, carved into a bunch of zvols, each exported via
2007 May 23
13
Preparing to compare Solaris/ZFS and FreeBSD/ZFS performance.
Hi. I''m all set for doing performance comparsion between Solaris/ZFS and FreeBSD/ZFS. I spend last few weeks on FreeBSD/ZFS optimizations and I think I''m ready. The machine is 1xQuad-core DELL PowerEdge 1950, 2GB RAM, 15 x 74GB-FC-10K accesses via 2x2Gbit FC links. Unfortunately the links to disks are the bottleneck, so I''m going to use not more than 4 disks, probably.
2010 Mar 29
19
sharing a ssd between rpool and l2arc
Hi, as Richard Elling wrote earlier: "For more background, low-cost SSDs intended for the boot market are perfect candidates. Take a X-25V @ 40GB and use 15-20 GB for root and the rest for an L2ARC. For small form factor machines or machines with max capacity of 8GB of RAM (a typical home system) this can make a pleasant improvement over a HDD-only implementation." For the upcoming
2010 Jun 24
1
Gonna be stupid here...
But it''s early (for me), and I can''t remember the answer here. I''m sizing an Oracle database appliance. I''d like to get one of the F20 96GB flash accellerators to play with, but I can''t imagine I''d be using the whole thing for ZIL. The DB is likely to be a couple TB in size. Couple of questions: (a) since everything is going to be
2007 Mar 06
16
2007/128 SMF services for Xen
I am sponsoring this fasttrack for John Levon. It is set to expire on 3/14/2007. Note that this is an externally visible case. liane --- SMF services for Xen 1. Introduction This case introduces the SMF services used by a Solaris-based domain 0 when running on Xen, or a Xen-compatible hypervisor. All of these services only run on domain 0 when booted under Xen virtualisation.
2007 Jan 08
11
NFS and ZFS, a fine combination
Just posted: http://blogs.sun.com/roch/entry/nfs_and_zfs_a_fine ____________________________________________________________________________________ Performance, Availability & Architecture Engineering Roch Bourbonnais Sun Microsystems, Icnc-Grenoble Senior Performance Analyst 180, Avenue De L''Europe, 38330, Montbonnot Saint
2007 Jan 22
5
Final Xen SMF fast-track
This fast-track covers the SMF-related changes for dom0. This is a final draft, so any comments by the end of the week please... Man pages are attached. thanks john SMF services for Xen 1. Introduction This case introduces the SMF services used by a Solaris-based domain 0 when running on Xen, or a Xen-compatible hypervisor. All of these services only run on domain 0 when booted
2010 Jan 12
11
How do separate ZFS filesystems affect performance?
I''m working with a Cyrus IMAP server running on a T2000 box under Solaris 10 10/09 with current patches. Mailboxes reside on six ZFS filesystems, each containing about 200 gigabytes of data. These are part of a single zpool built on four Iscsi devices from our Netapp filer. One of these ZFS filesystems contains a number of global and per-user databases in addition to one sixth of the
2012 Nov 22
19
ZFS Appliance as a general-purpose server question
A customer is looking to replace or augment their Sun Thumper with a ZFS appliance like 7320. However, the Thumper was used not only as a protocol storage server (home dirs, files, backups over NFS/CIFS/Rsync), but also as a general-purpose server with unpredictably-big-data programs running directly on it (such as corporate databases, Alfresco for intellectual document storage, etc.) in order to
2011 Jul 15
22
Zil on multiple usb keys
This might be a stupid question, but here goes... Would adding, say, 4 4 or 8gb usb keys as a zil make enough of a difference for writes on an iscsi shared vol? I am finding reads are not too bad (40is mb/s over gige on 2 500gb drives stripped) but writes top out at about 10 and drop a lot lower... If I where to add a couple usb keys for zil, would it make a difference? Thanks. Sent from a
2010 Oct 06
14
Bursty writes - why?
I have a 24 x 1TB system being used as an NFS file server. Seagate SAS disks connected via an LSI 9211-8i SAS controller, disk layout 2 x 11 disk RAIDZ2 + 2 spares. I am using 2 x DDR Drive X1s as the ZIL. When we write anything to it, the writes are always very bursty like this: ool 488K 20.0T 0 0 0 0 xpool 488K 20.0T 0 0 0 0 xpool
2008 Jun 02
6
[Bug 2114] New: delegation_004: a non-root user can''t do ''zfs key -c'' with keychange delegated
http://defect.opensolaris.org/bz/show_bug.cgi?id=2114 Summary: delegation_004: a non-root user can''t do ''zfs key -c'' with keychange delegated Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: normal
2008 May 26
5
[Bug 2033] New: ''zfs create'' causes panic if key file doesn''t exist
http://defect.opensolaris.org/bz/show_bug.cgi?id=2033 Summary: ''zfs create'' causes panic if key file doesn''t exist Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: minor Priority: P2 Component:
2012 Aug 08
2
[LLVMdev] VLIW code generation for LLVM backend
Larin, Thank you for telling me about this. Our lab is planning to design a VLIW DSP and has to make a choice between GCC and LLVM, for which I take responsibility. As we all know that GCC's codes possess a long history and has a somewhat bad learning curve, I suggest choosing LLVM. It seems now the only drawback is its poor support for VLIW architecture. And so if we can count on
2007 Oct 11
3
CVS/SVN conversion to GIT finished for Samba 3.x branches
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Folks, The migration is done and the new git repos are online. You can browse the repositories as well as find links to the documentation in the Samba wiki at http://gitweb.samba.org/ A few things, such as buildfarm integration, are still underway and will be completed in a few days. cheers, jerry - --
2004 Dec 04
5
Is Gigabit Ethernet necessary?
For an office that is using VoIP phones to connect to Asterisk, is gigabit ethernet really necessary for the Asterisk box to connect to the switch? I know that I won't even approach the limits of 100 Mbps, but would gigabit help with latency / collisions when several calls are underway? The fact is, anything going outside the office will be over a data T1, so intuition tells me that 100
2008 Feb 25
3
[Bug 631] New: zpool get with no pool name dumps core in zfs-crypto
http://defect.opensolaris.org/bz/show_bug.cgi?id=631 Summary: zpool get with no pool name dumps core in zfs-crypto Classification: Development Product: zfs-crypto Version: unspecified Platform: Other OS/Version: Solaris Status: NEW Severity: minor Priority: P4 Component: other AssignedTo: