similar to: Bugs fixed in update 4?

Displaying 20 results from an estimated 30000 matches similar to: "Bugs fixed in update 4?"

2007 Sep 19
8
ZFS Solaris 10u5 Proposed Changes
ZFS Fans, Here''s a list of features that we are proposing for Solaris 10u5. Keep in mind that this is subject to change. Features: PSARC 2007/142 zfs rename -r PSARC 2007/171 ZFS Separate Intent Log PSARC 2007/197 ZFS hotplug PSARC 2007/199 zfs {create,clone,rename} -p PSARC 2007/283 FMA for ZFS Phase 2 PSARC/2006/465 ZFS Delegated Administration PSARC/2006/577 zpool property to
2007 Sep 19
7
ZFS Solaris 10 Update 4 Patches
The latest ZFS patches for Solaris 10 are now available: 120011-14 - SunOS 5.10: kernel patch 120012-14 - SunOS 5.10_x86: kernel patch ZFS Pool Version available with patches = 4 These patches will provide access to all of the latest features and bug fixes: Features: PSARC 2006/288 zpool history PSARC 2006/308 zfs list sort option PSARC 2006/479 zfs receive -F PSARC 2006/486 ZFS canmount
2007 Apr 19
14
Permanently removing vdevs from a pool
Is it possible to gracefully and permanently remove a vdev from a pool without data loss? The type of pool in question here is a simple pool without redundancies (i.e. JBOD). The documentation mentions for instance offlining, but without going into the end results of doing that. The thing I''m looking for is an option to evacuate, for the lack of a better word, the data from a specific
2007 Jul 04
3
zfs dynamic lun expansion
Hi, I had 2 luns in a zfs mirrored config. I increased the size of both the luns by x gig and offlined/online the individual luns in the zpool, also tried an export/import of the zpool, but i am unable to see the increased size....what would i need to do to see the increased size?...or is it not possible yet? This message posted from opensolaris.org
2007 Jun 01
10
SMART
On Solaris x86, does zpool (or anything) support PATA (or SATA) IDE SMART data? With the Predictive Self Healing feature, I assumed that Solaris would have at least some SMART support, but what I''ve googled so far has been discouraging. http://prefetch.net/blog/index.php/2006/10/29/solaris-needs-smart-support-please-help/ Bug ID: 4665068 SMART support in IDE driver
2008 Mar 26
7
Lustre Solution Delivery
What is the process for integrating a Lustre+SAMFS solution into an existing customer environment. The plan is to have CRS build the Lustre component, but Lustre and SAMFS will need to configured and integrated into the customer computing environment. I am very familiar with the SAMFS integration, but not Lustre integration. Do we have resources in PS to provide the integration? Is this
2007 Sep 26
2
TCP connections not getting cleaned up after application exits
Howdy, We are running zones on a number of Solaris 10 update 3 hosts, and we are bumping into an issue where the kernel doesn''t clean up connections after an application exits. When this issue occurs, the netstat utility doesn''t show anything listening on the port the application uses (8080 in the example below), but connections are still listed in the ESTABLISHED state: $
2007 Sep 18
5
ZFS panic in space_map.c line 125
One of our Solaris 10 update 3 servers paniced today with the following error: Sep 18 00:34:53 m2000ef savecore: [ID 570001 auth.error] reboot after panic: assertion failed: ss != NULL, file: ../../common/fs/zfs/space_map.c, line: 125 The server saved a core file, and the resulting backtrace is listed below: $ mdb unix.0 vmcore.0 > $c vpanic() 0xfffffffffb9b49f3() space_map_remove+0x239()
2006 Apr 06
4
Why is my kernel eating my memory
Can someone, more learned in the ways of dtrace point me at what to look at to help understand why the kernel on one machine is using tons of memory, while another machine doing the same task/same user load is not. swapinfo for the "afflicted" machine shows RAM _______Total 16384 Mb RAM Unusable 73 Mb RAM Kernel 9226 Mb RAM Locked 2 Mb RAM Used
2012 Nov 13
9
Intel DC S3700
[This email is either empty or too large to be displayed at this time]
2017 Sep 08
1
cyrus spool on btrfs?
I think it depends on who you ask. Facebook and Netflix are using it extensively in production: https://www.linux.com/news/learn/intro-to-linux/how-facebook-uses-linux-and-btrfs-interview-chris-mason Though they have the in-house kernel engineering resources to troubleshoot problems. When I see quotes like this [1] on the product's WIKI: "The parity RAID code has multiple serious
2006 Jul 17
28
Big JBOD: what would you do?
ZFS fans, I''m preparing some analyses on RAS for large JBOD systems such as the Sun Fire X4500 (aka Thumper). Since there are zillions of possible permutations, I need to limit the analyses to some common or desirable scenarios. Naturally, I''d like your opinions. I''ve already got a few scenarios in analysis, and I don''t want to spoil the brain storming, so
2010 Nov 05
3
ZFS vs mpxio vs cfgadm in Solaris.
Folks, I''m trying to figure out whether we should give ZFS / mpxio a shot on one of our research servers, or simply skip it (as we have previously). In Nov 2009 Cindy responded to a thread concerning ZFS device issues, cfgadm, and mpxio: http://mail.opensolaris.org/pipermail/zfs-discuss/2009-November/033496.html I''ve got an x2270 with the Sun EZ-SAS HBA and external SATA
2008 Apr 15
1
Flag day: ZFS Boot Support
On Fri, Apr 11, 2008 at 06:39:00PM -0700, George Wilson wrote: > Solaris Installation Guide > System Administration Guide: Basic > ZFS Administration Guide > System Administration Guide: Devices and File Systems Where can I get the updated guides? > For further information, visit our OpenSolaris page at: > >
2006 Mar 03
3
Using DTrace to locate memory leak
Howdy, I am attempting to isolate the location of a memory leak in a 4GL program, and have hit a bit of a snag. When I LD_PRELOAD libumem and run the application server, "::findleaks -fdv" reports numerous leaks: CACHE LEAKED BUFCTL CALLER 0000000100b49068 15 0000000100ef2d50 fdcon+0x6c4 0000000100b50028 1 0000000100b72ac0 fdcon+0x6c4 0000000100b49068
2009 Apr 19
21
[on-discuss] Reliability at power failure?
Casper.Dik at Sun.COM wrote: > > I would suggest that you follow my recipe: not check the boot-archive > during a reboot. And then report back. (I''m assuming that that will take > several weeks) > We are back at square one; or, at the subject line. I did a zpool status -v, everything was hunky dory. Next, a power failure, 2 hours later, and this is what zpool status
2016 Mar 25
2
Dovecot on FreeBSD 10.1
I'm experiencing the same problem as mentioned earlier in the mailing list. - Is there some kind of workaround for this issue? http://dovecot.org/pipermail/dovecot/2015-April/100549.html /S?ren
2010 Aug 13
32
ZFS development moving behind closed doors
If this information is correct, http://opensolaris.org/jive/thread.jspa?threadID=133043 further development of ZFS will take place behind closed doors. Opensolaris will become the internal development version of Solaris with no public distributions. The community has been abandoned. -- -Gary Mills- -Unix Group- -Computer and Network Services-
2016 Mar 25
3
Dovecot on FreeBSD 10.1
Hi Larry, I?m running dovecot2-2.2.22. ? I?ve tried to compile dovecot from the souce, and I see the same error :/ Are you using sql auth in your setup? root at beastie:/var/log # doveconf -n # 2.2.22 (fe789d2): /usr/local/etc/dovecot/dovecot.conf # Pigeonhole version 0.4.13 (7b14904) # OS: FreeBSD 10.1-RELEASE-p6 amd64 zfs auth_debug = yes auth_debug_passwords = yes auth_mechanisms = plain
2009 Dec 24
6
benchmark results
I've had the chance to use a testsystem here and couldn't resist running a few benchmark programs on them: bonnie++, tiobench, dbench and a few generic ones (cp/rm/tar/etc...) on ext{234}, btrfs, jfs, ufs, xfs, zfs. All with standard mkfs/mount options and +noatime for all of them. Here are the results, no graphs - sorry: http://nerdbynature.de/benchmarks/v40z/2009-12-22/ Reiserfs