similar to: How to properly read "zpool iostat -v" ? ;)

Displaying 20 results from an estimated 90 matches similar to: "How to properly read "zpool iostat -v" ? ;)"

2011 Nov 25
1
Recovering from kernel panic / reboot cycle importing pool.
Yesterday morning I awoke to alerts from my SAN that one of my OS disks was faulty, FMA said it was in hardware failure. By the time I got to work (1.5 hours after the email) ALL of my pools were in a degraded state, and "tank" my primary pool had kicked in two hot spares because it was so discombobulated. ------------------- EMAIL ------------------- List of faulty resources:
2007 May 09
1
dsl iso and xen
Hi Im trying to get DSL(damm small linux) to run in Xen, DSL is a small live-iso with linux. I have made a simple configuration file, se below: <configuration file> kernel = "/boot/vmlinuz-xen" memory = 64 name = ''dsl'' disk = [''file:/tmp/dsl-3.3.iso,sda2,w''] root = "/dev/sda2 ro" <configuration file>
2008 Jun 30
0
[PATCH] qemu xen-console, limit buffering
Running this script from within a PV guest without having an xm console attached can use all dom0 memory. while [ 1 ]; do ls -lR /usr > /dev/xvc0; done Start of script: (1% mem usage) PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 10281 root 15 0 76536 14m 2448 S 2 1.0 0:00.61 qemu-dm Couple minutes later: (17% mem usage)
2007 Apr 14
1
Move data from the zpool (root) to a zfs file system
Hi List, As a ZFS newbie, I foolishly copied my data set to the root zpool file system (a large iSCSI SAN array). Thus: # zpool create -f iscsi c4t19d0 c4t20d0 c4t21d0 c4t22d0 c4t23d0 c4t24d0 # zpool list NAME SIZE USED AVAIL CAP HEALTH ALTROOT iscsi 9.53T 64.5K 5.34T 0% ONLINE - # zfs set mountpoint=/mydisks/iscsi iscsi Then copied
2006 Oct 31
3
zfs: zvols minor #''s changing and causing probs w/ volumes
Team, **Please respond to me and my coworker listed in the Cc, since neither one of us are on this alias** QUICK PROBLEM DESCRIPTION: Cu created a dataset which contains all the zvols for a particular zone. The zone is then given access to all the zvols in the dataset using a match statement in the zoneconfig (see long problem description for details). After the initial boot of the zone
2015 Nov 04
0
getting a CentOS6 VM on VMware ESXi platform to recognize a new disk device
>> > vmware esxi 5.5.0 (free, using vsphere client to manage), vm is minimal > centos 7 64bit. I added a 16gb vdisk and immediately see this in dmesg... > > [155484.386792] vmw_pvscsi: msg type: 0x0 - MSG RING: 1/0 (5) > [155484.386796] vmw_pvscsi: msg: device added at scsi0:1:0 > [155484.388250] scsi 0:0:1:0: Direct-Access VMware Virtual disk > 1.0 PQ: 0 ANSI:
2015 Nov 04
2
getting a CentOS6 VM on VMware ESXi platform to recognize a new disk device
On 11/4/2015 11:36 AM, Boris Epstein wrote: > Absolutely, I see your point. This was the starting point - you add the > device on the ESXi server, you reboot the VM, the VM sees the device, no > problem. Now, I ask - do I have to reboot the VM? Logically I hope there > ought to be a way for me not to have to do that - but I have yet to figure > out how to get there. > vmware
2008 Aug 27
2
Bad file descriptor with maildir and bzip2 files
Hi, I'm using the zlib plugin with dovecot 1.1.2 (on Centos 4.5, x86, ext3) and some bzip2 (version 1.0.2) compressed mail files are causing the following error to show in the logs when I click on the folder in outlook express. Aug 27 11:39:15 server-2 dovecot: IMAP(marchive): fstat(maildir) failed: Bad file descriptor Aug 27 11:39:17 Info: IMAP(marchive): Loading modules from directory:
2012 Jun 13
0
ZFS NFS service hanging on Sunday morning problem
> > Shot in the dark here: > > What are you using for the sharenfs value on the ZFS filesystem? Something like rw=.mydomain.lan ? They are IP blocks or hosts specified as FQDNs, eg., pptank/home/tcrane sharenfs rw=@192.168.101/24,rw=serverX.xx.rhul.ac.uk:serverY.xx.rhul.ac.uk > > I''ve had issues where a ZFS server loses connectivity to the primary DNS server and
2006 Oct 31
0
6269165 misleading comments in usr/src/cmd/stat/iostat/iostat.c
Author: petede Repository: /hg/zfs-crypto/gate Revision: ef5d30f0abff39691966b9805b895f3c757b5a9f Log message: 6269165 misleading comments in usr/src/cmd/stat/iostat/iostat.c 6286482 remove the only occurrence of "shit" in OpenSolaris 6231501 Typo in <sys/fem.h> Contributed by Shawn Walker <binarycrusader at gmail.com>. Files: update: usr/src/cmd/stat/iostat/iostat.c
2019 Aug 30
0
NFS Server on Centos 7.6.1810 dying without reason
I managed to get NFS statistics just before server rebooted. I can reproduce this every time. Question: Am I saturating the SATA bus on the NFS server so violently with the SSDs that the server hard reboots? Perhaps due to the RAID 6 nature of operations? Server info AMD EPYC 7281 16CORE 32 GIG RAM 6xSATA3 SSD 1.9TB in mdadm RAID6 OS: Centos 7.6 SATA controller: Advanced Micro Devices, Inc.
2008 Jan 03
1
The iostat command
Hi All, I am learning iostat command to understand disk I/O statistics. We have 2 Centos 4 servers running where oracle is installed.We installed them 2 weeks ago. @ that time, These Servers performed well. But, Now We have come to know that these 2 Machines are quite slow when comparing to the first week. So, Some say, run iostat to see statictics. I am not familiar with comamnd well. I
2008 Apr 29
0
zpool attach vs. zpool iostat
Hello zfs-discuss, S10U4+patches, SPARC If I attach a disk to vdev in a pool to get mirrored configuration then during resilver zpool iostat 1 will report only reads being done from pool and basically no writes. If I do zpool iostat -v 1 then I can see it is writing to new device however on a pool and mirror/vdev level it is still reporting only reads. If during resilvering reads
2007 Apr 15
0
zpool iostat : This command can be tricky ...
I really need to take a longer look here. /* * zpool iostat [-v] [pool] ... [interval [count]] * * -v Display statistics for individual vdevs * * This command can be tricky because we want to be able to deal with pool . . . I think I may need to deal with a raw option here ? /* * Enter the main iostat loop. */ cb.cb_list = list;
2010 May 28
0
zpool iostat question
Following is the output of "zpool iostat -v". My question is regarding the datapool row and the raidz2 row statistics. The datapool row statistic "write bandwidth" is 381 which I assume takes into account all the disks - although it doesn''t look like it''s an average. The raidz2 row static "write bandwidth" is 36, which is where I am confused. What
2009 Jan 30
1
RFE: parsable iostat and zpool layout
I would like zpool iostat to take a "-p" option to output parsable statistics with absolute counters/figures that for example could be fed to MRTG, RRD, et al. The "zpool iostat [-v] POOL 60 [N]" is great for humans but not very api-friendly; N=2 is a bit overkill and unreliable. Is this info available in kstat, or is this an RFE candidate? In Solaris10? Ditto for zpool
2006 May 23
1
iostat numbers for ZFS disks, build 39
I updated an i386 system to b39 yesterday, and noticed this when running iostat: r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 0.0 0.5 0.0 10.0 0.0 0.0 0.0 0.5 0 0 c0t0d0 0.0 0.5 0.0 10.0 0.0 0.0 0.0 0.6 0 0 c0t1d0 0.0 65.1 0.0 119640001.5 0.0 0.0 0.0 0.3 0 2 c0t2d0 0.0 65.1 0.0 119640090.2 0.0
2009 Jan 12
3
Top and Iostat
Is there anyway to tell with top or iostat which process is hogging all the disk I/O? Matt
2010 Mar 11
1
zpool iostat / how to tell if your iop bound
What is the best way to tell if your bound by the number of individual operations per second / random io? "zpool iostat" has an "operations" column but this doesn''t really tell me if my disks are saturated. Traditional "iostat" doesn''t seem to be the greatest place to look when utilizing zfs. Thanks, Chris -------------- next part -------------- An
2009 Sep 21
2
Question about iostat output
Hello, We are planning to moving most of our servers to ESX but before buying our SAN, we want to do some I/O stats to see if iSCSI is enough or if we have to go with FC. So I found a plugin for Nagios that can log I/O stats with iostat. So far it's fine with single disk/one partition servers, but on our Oracle Database 10g server, we have two drives in RAID 1 (/dev/sda) and 4 other