Displaying 20 results from an estimated 80 matches similar to: "Solaris 10 11/06"
2012 Jan 11
3
Unable to allocate dma memory for extra SGL
Hi all;
We have a Solaris 10 U9 x86 instance running on Silicon Mechanics /
SuperMicro hardware.
Occasionally under high load (ZFS scrub for example), the box becomes
non-responsive (it continues to respond to ping but nothing else works
-- not even the local console). Our only solution is to hard reset
after which everything comes up normally.
Logs are showing the following:
Jan 8
2007 Sep 19
7
ZFS Solaris 10 Update 4 Patches
The latest ZFS patches for Solaris 10 are now available:
120011-14 - SunOS 5.10: kernel patch
120012-14 - SunOS 5.10_x86: kernel patch
ZFS Pool Version available with patches = 4
These patches will provide access to all of the latest features and bug
fixes:
Features:
PSARC 2006/288 zpool history
PSARC 2006/308 zfs list sort option
PSARC 2006/479 zfs receive -F
PSARC 2006/486 ZFS canmount
2009 Jan 24
0
fsstat Provider
I''ve been playing with the fsstat provider lately. I''m curious why its
gone undocumented and largely ignored (per google). I admit that I did
experience some oddities with the data, so I''d also like to rule out
whether or not it has problems which might have caused it to be left in
quietly.
benr.
2008 Jul 05
4
iostat and monitoring
Hi gurus,
I like zpool iostat and I like system monitoring, so I setup a script
within sma to let me get the zpool iostat figures through snmp.
The problem is that as zpool iostat is only run once for each snmp
query, it always reports a static set of figures, like so:
root at exodus:snmp # zpool iostat -v
capacity operations bandwidth
pool used avail read
2010 Mar 06
3
Monitoring my disk activity
Recently, I''m benchmarking all kinds of stuff on my systems. And one
question I can''t intelligently answer is what blocksize I should use in
these tests.
I assume there is something which monitors present disk activity, that I
could run on my production servers, to give me some statistics of the block
sizes that the users are actually performing on the production server.
2012 Jun 06
24
Occasional storm of xcalls on segkmem_zio_free
So I have this dual 16-core Opteron Dell R715 with 128G of RAM attached
to a SuperMicro disk enclosure with 45 2TB Toshiba SAS drives (via two
LSI 9200 controllers and MPxIO) running OpenIndiana 151a4 and I''m
occasionally seeing a storm of xcalls on one of the 32 VCPUs (>100000
xcalls a second). The machine is pretty much idle, only receiving a
bunch of multicast video streams and
2006 May 24
1
New features in Solaris Express 05/06
Will the ability to import a destroyed ZFS pool and the fsstat command that''s part of the latest Solaris Express release (B38) make it into Solaris 10 Update-2 when it''s released in June/July??? Also has any decision been made yet what build Update-2 will be taken from to give an idea of what can be expected for ZFS???
This message posted from opensolaris.org
2006 Apr 06
0
NFSv3 and File operation scripts
Howdy,
I wrote a pair of scripts to measure file and NFSv3 operations, and
thought I would share them with the folks on the list. You can view
the script output by pointing your browser at the following URLs:
Per Process NFSv3 Client statistics (inspired by fsstat/nfsstat):
http://daemons.net/~matty/code/nfsclientstats.pl.txt
Per Process File Operations (inspired by fsstat):
2009 Apr 12
7
Any news on ZFS bug 6535172?
We''re running a Cyrus IMAP server on a T2000 under Solaris 10 with
about 1 TB of mailboxes on ZFS filesystems. Recently, when under
load, we''ve had incidents where IMAP operations became very slow. The
general symptoms are that the number of imapd, pop3d, and lmtpd
processes increases, the CPU load average increases, but the ZFS I/O
bandwidth decreases. At the same time, ZFS
2007 Oct 27
14
X4500 device disconnect problem persists
After applying 125205-07 on two X4500 machines running Sol10U4 and
removing "set sata:sata_func_enable = 0x5" from /etc/system to
re-enable NCQ, I am again observing drive disconnect error messages.
This in spite of the patch description which claims multiple fixes
in this area:
6587133 repeated DMA command timeouts and device resets on x4500
6538627 x4500 message logs contain multiple
2012 Jul 04
1
dovecot and nfs readdir vs readdirplus operations
Hello,
We are having performance problems trying to migrate our pop/imap
servers to a new version. Our old servers are 4 debian lenny with 5GB of
RAM running of XenServer VMs with kernel 2.6.32-4-amd64 and dovecot
1.1.16. New servers are 4 ubuntu 12.04 with dovecot 2.1.5 running on
vmware vm with 6 cores and 16GB of RAM and kernel 3.2.0-24-generic. On
both server we are using nfs 3 with
2010 Jan 08
0
ZFS partially hangs when removing an rpool mirrored disk while having some IO on another pool on another partition of the same disk
Hello,
Sorry for the (very) long subject but I''ve pinpointed the problem to this exact situation.
I know about the other threads related to hangs, but in my case there was no < zfs destroy > involved, nor any compression or deduplication.
To make a long story short, when
- a disk contains 2 partitions (p1=32GB, p2=1800 GB) and
- p1 is used as part of a zfs mirror of rpool
2004 Jul 06
0
destroyed files using shares on nfs-mounted filesystem
Hallo,
we are using samba V.3.0.2a on Linux 2.4.18
Previous Versions of samba showed the same effect.
The linux-box has mounted nfs-shares from solaris SunOS 5.8
This nfs-connection is a WAN-connection, that has limited bandwidth(some Megabit/second). When connection is slow, then it happened often that saving a file to this share is disturbed. Sometime it simply hangs, sometime an error
2006 Jul 21
1
LDA Command time limit exceeded
I have many errors of this type in my logs:
2 * postfix/pipe[21511]: BC8DD3356A: to=<*****@******>, relay=dovecot,
delay=1001, status=bounced (Command time limit exceeded:
"/usr/local/dovecot-1.0-rc2/libexec/dovecot/deliver")
My command time limit is 1000s in postfix (more than 15 minutes)
After many check, I found that a lots of this error is not really an
error.
The mail is
2002 Sep 05
2
AIX & Large File Support Problem (+ Solution)
Hi,
I just wanted to relate the solution to a problem I was having to hopefully
save someone else a day of frustration. I'm using rsync-2.5.5 on AIX 4.3,
compiled with gcc 2.95.3. The file I was sync'ing was very large (>2GB).
Despite being configured with --enable-largefiles (which #defines
_LARGE_FILES properly for AIX), and despite the fact that the initial
transfer of said file
2018 Jan 08
0
Re: virtdf outputs on host differs from df in guest
The answer is attached. In summary it's not a bug in virt-df,
nor df, nor the kernel.
Rich.
--
Richard Jones, Virtualization Group, Red Hat http://people.redhat.com/~rjones
Read my programming and virtualization blog: http://rwmj.wordpress.com
virt-df lists disk usage of guests without needing to install any
software inside the virtual machine. Supports Linux and Windows.
2018 Jan 07
3
Re: virtdf outputs on host differs from df in guest
after install libguestfs_xfs, all results are:
[using guestfish]
guestfish -N fs:xfs -m /dev/sda1 statvfs /
bsize: 4096
frsize: 4096
blocks: 24713
bfree: 23391
bavail: 23391
files: 51136
ffree: 51133
favail: 51133
fsid: 2049
flag: 4096
namemax: 255
[using virt-rescure]
virt-rescue -a test1.img
><rescue> mount /dev/sda1 /sysroot
><rescue> stat -f /sysroot
File:
2008 Dec 05
0
resync onnv_105 partial for 6713916
Author: Darren Moffat <Darren.Moffat at Sun.COM>
Repository: /hg/zfs-crypto/gate
Latest revision: 957d30a3607ed9f3cbe490da5894d1e1b2104033
Total changesets: 28
Log message:
resync onnv_105 partial for 6713916
Files:
usr/src/Makefile.lint
usr/src/Targetdirs
usr/src/cmd/Makefile
usr/src/cmd/Makefile.cmd
usr/src/cmd/acctadm/Makefile
usr/src/cmd/acctadm/acctadm.xcl
2007 Apr 11
0
raidz2 another resilver problem
Hello zfs-discuss,
One of a disk started to behave strangely.
Apr 11 16:07:42 thumper-9.srv sata: [ID 801593 kern.notice] NOTICE: /pci at 1,0/pci1022,7458 at 3/pci11ab,11ab at 1:
Apr 11 16:07:42 thumper-9.srv port 6: device reset
Apr 11 16:07:42 thumper-9.srv scsi: [ID 107833 kern.warning] WARNING: /pci at 1,0/pci1022,7458 at 3/pci11ab,11ab at 1/disk at 6,0 (sd27):
Apr 11 16:07:42 thumper-9.srv
2006 Oct 31
0
6433264 crash when adding spare: nvlist_lookup_string(cnv, "path", &path) == 0
Author: eschrock
Repository: /hg/zfs-crypto/gate
Revision: 975fad4fa05753c23a18254797cce42bb33b4375
Log message:
6433264 crash when adding spare: nvlist_lookup_string(cnv, "path", &path) == 0
6433406 zfs_open() can leak memory on failure
6433408 namespace_reload() can leak memory on allocation failure
6433679 zpool_refresh_stats() has poor error semantics
6433680 changelist_gather()