similar to: disk utilization is over 200%

Displaying 20 results from an estimated 200 matches similar to: "disk utilization is over 200%"

2012 Jul 11
2
For loop help
Hi Everyone, I could use help developing a for loop. I have a dataset with tallies for a number of species within 7 different size classes. I need to uncollate the data into the rawest form (ie: each row a different individual) & retrain the metadata associated with each row (Date, Recorder, Site, Transect and Species). Here is an example of the current format:
2010 Oct 04
8
Can I "upgrade" a striped pool of vdevs to mirrored vdevs?
Hi, once I created a zpool of single vdevs not using mirroring of any kind. Now I wonder if it''s possible to add vdevs and mirror the currently existing ones. Thanks, budy -- This message posted from opensolaris.org
2002 Dec 25
1
some weird filename related error
Both OggDropXPd 1.6 beta 13 & the modified version of CDex from the site i got OggDrop XPd 1.6 beta 13, and the ogg output plugin for winamp have a problem with this filename: Pre Club Mix (Mixed by DJ Pete Gooding) - Cafe Mambo The Real Sound Of Ibiza - David Morales and Albert Cabrera present Moca feat Deanna - Higher (Mo and Al's Original Mix) and this one 001_Pre Club Mix (Mixed by
2009 Jul 09
1
Problem building 3.4.0 from source
I am attempting to build 3.4.0 from source on my Ubuntu 8.04 box. I have gcc installed. When I run ./configure it returns "C compiler cannot create executables". I have been looking through the config.log file and I think I found the error message. configure:3304: gcc conftest.c >&5 /usr/bin/ld: crt1.o: No such file: No such file or directory collect2: ld returned
2009 Jan 12
1
ZFS size is different ?
Hi all, I have 2 questions about ZFS. 1. I have create a snapshot in my pool1/data1, and zfs send/recv it to pool2/data2. but I found the USED in zfs list is different: NAME USED AVAIL REFER MOUNTPOINT pool2/data2 160G 1.44T 159G /pool2/data2 pool1/data 176G 638G 175G /pool1/data1 It keep about 30,000,000 files. The content of p_pool/p1 and backup/p_backup
2009 Dec 16
27
zfs hanging during reads
Hi, I hope there''s someone here who can possibly provide some assistance. I''ve had this read problem now for the past 2 months and just can''t get to the bottom of it. I have a home snv_111b server, with a zfs raid pool (4 x Samsung 750GB SATA drives). The motherboard is a ASUS M2N68-CM (4 SATA ports) with an Athlon LE1620 single core CPU and 4GB of RAM. I am using it
2007 Oct 17
3
Dtrace scripts for performance data gathering
I am looking for Dtrace scripts that can be used to collect data during performance tests. I am especially interested in IO but CPU, memory, threads, etc are needed as well. Thanks, Dave
2008 Jul 15
1
Cannot share RW, "Permission Denied" with sharenfs in ZFS
Hi everyone, I have just installed Solaris and have added a 3x500GB raidz drive array. I am able to use this pool (''tank'') successfully locally, but when I try to share it remotely, I can only read, I cannot execute or write. I didn''t do anything other than the default ''zfs set sharenfs=on tank''... how can I get it so that any allowed user can access
2008 Jul 15
2
Cannot share RW, "Permission Denied" with sharenfs in ZFS
Hi everyone, I have just installed Solaris and have added a 3x500GB raidz drive array. I am able to use this pool (''tank'') successfully locally, but when I try to share it remotely, I can only read, I cannot execute or write. I didn''t do anything other than the default ''zfs set sharenfs=on tank''... how can I get it so that any allowed user can access
2006 Apr 06
15
A few Newbie questions about RAIDZ
1. I have a 4x18GB drive setup as RAIDZ. Now when thinking about it in terms of RAID5 I would expect to get (4-1)x18 worth of drive space, but DF -h shows 4x18. Is this a bug or do I not understand? 2. Once again thinking in RAID5 terms if I have 4X18GB and 12X9GB drives and I want to make a RAIDZ of all of them I would expect the 18GB to be treated at 9GB so the RAIDZ would be 16X9GB. Is
2010 May 28
21
expand zfs for OpenSolaris running inside vm
hello, all I am have constraint disk space (only 8GB) while running os inside vm. Now i want to add more. It is easy to add for vm but how can i update fs in os? I cannot use autoexpand because it doesn''t implemented in my system: $ uname -a SunOS sopen 5.11 snv_111b i86pc i386 i86pc If it was 171 it would be grate, right? Doing following: o added new virtual HDD (it becomes
2009 Jul 09
6
Active Directory Integration Problems
Hello everyone, I have setup Samba 3.0.28a on an Ubuntu 8.04 server. The setup that I am working with is an exact copy (as far as I can tell) if an identical installation that I did on a test box. Kerberos is setup and working properly. I can use kinit to issue tickets. The box has been successfully joined to the Active Directory domain. I can enumerate AD users and groups. I can log
2008 Nov 05
2
plockstat: processing aborted: Abort due to systemic unresponsiveness
Hello, I need help here about plockstat on X86 platform (Sun X4600 AMD) # plockstat -A -p 20034 plockstat: processing aborted: Abort due to systemic unresponsiveness # plockstat -e 5 -s 10 -A -x bufsize=100k -x aggsize=20m -p 20034 plockstat: processing aborted: Abort due to systemic unresponsiveness # ps -ef | grep 20034 algodev 20034 1 2 07:00:54 ? 86:17
2012 Feb 16
2
help with e+01 number abbreviations
Dear List, I will appreciate any advice regarding how to convert the following numbers [I got in return by taxondive()] in numeric integers without the e.g. 6.4836e+01 abbreviations. Thank you very much in advance, Gian > taxa_dive Species Delta Delta* Lambda+ Delta+ S Delta+ Nat1 5.0000e+00 6.4836e+01 9.5412e+01 6.7753e+02 8.7398e+01 436.99 Nat2
2005 Nov 18
6
expected system load from DTrace scripts/probes
Does anyone out there have any thoughts on the type of load common DTrace scripts would cause on a system if run 24x7? I know "common DTrace scripts" and their underlining probe calls a vague statement. So for the lack of a common and establish set of scripts in the OS, I''ll use the most popular right for my question...the DTraceToolkit from Brendan Gregg. Which by the
2007 Dec 27
2
Help with lm and multiple linear regression? (Plain Text version)
(Apologies the previous version was sent as rich text) Hello, I'm new to R, but I've read the intro to R and successfully connected it to an instance of mysql. I'm trying to perform multiple linear regression, but I'm having trouble using the lm function. To start, I have read in a simply y matrix of values(dependent variable) and x matrix of independent variables. It says both
2007 Jan 10
0
[DTrace] %CPU, %Mem, %Disk and %Net by PID
G''Day Folks, Well, this one was my white whale. I now have a demonstration program that uses Perl, Kstat, /proc/*/usage, /proc/*/psinfo, and DTrace. The program itself is more of a prototype, I want to get serious about it and rewrite it in C. It currently has output like, # prustat -t5 -snet 10 PID %CPU %Mem %Disk %Net COMM 90 0.00 0.09 0.00 0.00
2008 Apr 18
2
plockstat: failed to add to aggregate: Abort due to drop
when check java process lock statistics, plockstat failed, please see below: # prstat -mLp 21162 PID USERNAME USR SYS TRP TFL DFL LCK SLP LAT VCX ICX SCL SIG PROCESS/LWPID 21162 7677 0.9 0.1 0.0 0.0 0.0 99 0.0 0.3 83 89 215 0 java/81 21162 7677 0.3 0.1 0.0 0.0 0.0 0.0 99 0.2 106 33 305 0 java/35 21162 7677 0.1 0.0 0.0 0.0 0.0 100 0.0 0.1 79 6 85 0 java/59
2009 Aug 13
1
[PATCH] Don't assume grub is on a separate boot filesystem
Paths in grub.conf are relative to the filesystem containing it. grub parsing currently assumes that it is on /boot, and will fail if it isn't, for example because a guest only has a single partition. This patch makes grub parsing work harder to work out what grub paths are relative to. Firstly, it looks for a previous detected 'linux-grub' filesystem. If this isn't found, it
2017 Dec 06
1
reminder: Community meeting at 15:00 UTC in #gluster-meeting on Freenode IRC
Hi all, This a reminder about today's meeting. The meeting will start later today at 15:00 UTC. You can convert that to a local time with the following command in your terminal: $ date -d '15:00 UTC' Because the meeting is an open floor, topics and updates need to be added to the meeting pad at: https://bit.ly/gluster-community-meetings Hope to see you later online, Niels