similar to: high read iops - more memory for arc?

Displaying 20 results from an estimated 400 matches similar to: "high read iops - more memory for arc?"

2009 Dec 28
0
[storage-discuss] high read iops - more memory for arc?
Pre-fletching on the file and device level has been disabled yielding good results so far. We''ve lowered the number of concurrent ios from 35 to 1 causing the service times to go even lower (1 -> 8ms) but inflating actv (.4 -> 2ms). I''ve followed your recommendation in setting primarycache to metadata. I''ll have to check with our tester in the morning if it made
2010 Sep 29
10
Resliver making the system unresponsive
This must be resliver day :) I just had a drive failure. The hot spare kicked in, and access to the pool over NFS was effectively zero for about 45 minutes. Currently the pool is still reslivering, but for some reason I can access the file system now. Resliver speed has been beaten to death I know, but is there a way to avoid this? For example, is more enterprisy hardware less susceptible to
2010 Aug 19
0
Unable to mount legacy pool in to zone
So i''ve inherited this solaris 10 system with a sun storedge attached. pool: tol-pool state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM tol-pool ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c1t8d0 ONLINE 0 0 0 c1t9d0 ONLINE 0 0 0
2007 Jul 31
0
controller number mismatch
Hi, I just noticed something interesting ... don''t know whether it''s relevant or not (two commands run in succession during a ''nightly'' run): $ iostat -xnz 6 [...] extended device statistics r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device 0.0 0.3 0.0 0.8 0.0 0.0 0.2 0.2 0 0 c2t0d0 2.2
2008 Jul 11
3
Linux equivalent of 'format' in solaris
It shows the physical disks on the server bash-2.05b# format Searching for disks...done AVAILABLE DISK SELECTIONS: 0. c0t2d0 <SUN18G cyl 7506 alt 2 hd 19 sec 248> /sbus at 3,0/SUNW,fas at 3,8800000/sd at 2,0 1. c0t3d0 <SUN18G cyl 7506 alt 2 hd 19 sec 248> /sbus at 3,0/SUNW,fas at 3,8800000/sd at 3,0 2. c0t4d0 <SUN18G cyl 7506 alt 2
2007 Jan 10
4
[osol-discuss] Re: bare metal ZFS ? How To ?
this is off list on purpose ? > run zpool import, it will search all attached storage and give you a list > of availible pools. then run zpool import poolname or add a -f if you > didn''t export before the install/upgrade. assume worst case someone walks up to you and drops an array on you. They say "its ZFS an'' I need that der stuff ''k? " all
2007 Jan 09
2
ZFS Hot Spare Behavior
I physically removed a disk (c3t8d0 used by ZFS ''pool01'') from a 3310 JBOD connected to a V210 running s10u3 (11/06) and ''zpool status'' reported this: # zpool status pool: pool01 state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Attach the
2010 Apr 02
0
ARC Tail
Greeting All Can any one help me figure out the size of the ARC tail, i.e, the portion in ARC that l2_feed thread is reading from before pages are evicted from ARC. Is the size of this tail proportional to total ARC size ? L2ARC device size ? is tunable ?????? your feed back is highly appreciated -- Abdullah Al-Dahlawi PhD Candidate George Washington University Department. Of
2011 Jan 24
0
ZFS/ARC consuming all memory on heavy reads (w/ dedup enabled)
Greetings Gentlemen, I''m currently testing a new setup for a ZFS based storage system with dedup enabled. The system is setup on OI 148, which seems quite stable w/ dedup enabled (compared to the OpenSolaris snv_136 build I used before). One issue I ran into, however, is quite baffling: With iozone set to 32 threads, ZFS''s ARC seems to consume all available memory, making
2007 Jul 18
1
Converting exisitng ZFS pool to MPxIO
We have a Sun v890, and I''m interested converting existing ZFS zpool from c#t#d# to MPxIO. % zpool status pool: data state: ONLINE status: ONLINE scrub: scrub completed with 0 errors on Sun Jul 15 10:58:33 2007 config: NAME STATE READ WRITE CKSUM data ONLINE 0 0 0 mirror ONLINE 0 0 0 c1t2d0
2010 Dec 21
5
relationship between ARC and page cache
One thing I''ve been confused about for a long time is the relationship between ZFS, the ARC, and the page cache. We have an application that''s a quasi-database. It reads files by mmap()ing them. (writes are done via write()). We''re talking 100TB of data in files that are 100k->50G in size (the files have headers to tell the app what segment to map, so mapped chunks
2006 Oct 19
1
Problem Reading from .txt
I apologize that I've asked a similar question before, but being new to R I don't think I did a very good job of formating the question. I've included a text file since the date set is somewhat large. What I have is a huge string of numbers in a text file. The numbers are all separated by comma's and the groups are separated by a semicolon. What I would like to do is read each
2012 Aug 03
2
how to identify values from a column of a dataframe, and insert them in other data.frame with the corresponding id?
Hello, I’d like to do next, see if you could help me please: I have a csv called “datuak” with a id called “calee_id” and a colunm called “poids”. I have another csv called “datuak2” with the same id called “calee_id”, (although there are “calee_id” that are in “datuak” but not in “datuak2” and inverse), and a column called “kg_totales” in which the values are repeteated for each calee_id
2006 Oct 05
1
Matrix input problem
Hi, Included is an R.script that came from a much large date set being read in to R from a .txt file. Why is it that the first line codes with an error, but the second line works fine? Is there some line length limit in R? This happens at random places all through my data. Any help would be appreciated. -------------------------------------------------------------------- Bill Wyatt
2014 Mar 20
1
Optimizing Gluster (gfapi) for high IOPS
Hey folks, We've been running VM's on qemu using a replicated gluster volume connecting using gfapi and things have been going well for the most part. Something we've noticed though is that we have problems with many concurrent disk operations and disk latency. The latency gets bad enough that the process eats the cpu and the entire machine stalls. The place where we've seen it
2014 Nov 18
1
Storage IOPs Calculation for Qmail Server
Dear DovecotORG, In my organization, we are about to implement Qmail Server. * The number of current users will be 800, in future it may increase upto 1200. * The number of concurrent users will be 300. I am the engineer to deploy the Qmail in Linux server. I need to tell the storage team on the IOPs requirement. I requested 8TB usable space for the mail storage (can
2004 Feb 26
0
Iops Vorbis player
Hi, does anyone have, or know much about, the Iops portable players? Are they better/worse than the players from iriver? I am looking for a USB-thumb-drive style of music player (that plays vorbis of course). >From what I have been able to gather from the Iops website, it suits this purpose, but the site is all in Korean. I was hoping to find someone who had actually purchased one of these
2013 Mar 18
2
Disk iops performance scalability
Hi, Seeing a drop-off in iops when more vcpu''s are added:- 3.8.2 kernel/xen-4.2.1/single domU/LVM backend/8GB RAM domU/2GB RAM dom0 dom0_max_vcpus=2 dom0_vcpus_pin domU 8 cores fio result 145k iops domU 10 cores fio result 99k iops domU 12 cores fio result 89k iops domU 14 cores fio result 81k iops ioping . -c 3 4096 bytes from . (ext4 /dev/xvda1): request=1 time=0.1 ms 4096 bytes
2009 Dec 04
2
measuring iops on linux - numbers make sense?
Hello, When approaching hosting providers for services, the first question many of them asked us was about the amount of IOPS the disk system should support. While we stress-tested our service, we recorded between 4000 and 6000 "merged io operations per second" as seen in "iostat -x" and collectd (varies between the different components of the system, we have a few such
2008 Jul 06
2
Measuring ZFS performance - IOPS and throughput
Can anybody tell me how to measure the raw performance of a new system I''m putting together? I''d like to know what it''s capable of in terms of IOPS and raw throughput to the disks. I''ve seen Richard''s raidoptimiser program, but I''ve only seen results for random read iops performance, and I''m particularly interested in write