similar to: Swap: typical rehash. Why?

Displaying 20 results from an estimated 3000 matches similar to: "Swap: typical rehash. Why?"

2013 Jul 23
2
How often is kernel "touching" swap partition?
Hello, I am going to buy new laptop with two drives, one SSD and one HDD and I want to place swap partition on HDD which will be most time unused and therefore spinned down. Problem is, that I don't know, if or how often is kernel touching swap space even if there is lot of free memory and thus spinning HDD up. Thank you in advance
2018 Feb 05
2
Very slow rsync to gluster volume UNLESS `ls` or `find` scan dir on gluster volume first
Thanks for the report Artem, Looks like the issue is about cache warming up. Specially, I suspect rsync doing a 'readdir(), stat(), file operations' loop, where as when a find or ls is issued, we get 'readdirp()' request, which contains the stat information along with entries, which also makes sure cache is up-to-date (at md-cache layer). Note that this is just a off-the memory
2008 Mar 19
1
Way to not page swap programs
Is there a way to tell linux "dont ever swap out my program"... Like perhaps a list of programs (some setup file) that if any program in my file listing is running dont consider the program when looking for something to swap out? Does anything like that exist? Thanks, Jerry
2009 Nov 16
1
Why swap if there's still physical memory available
Dear CentOS people, This is just a general question related to memory management, and there may have been a thread or two about it before, but I'd like to post anyway. A user was looking at top, whereby he found out that his two processes were 10.2g and 4836m in VIRT. They were 6.4g and 4.6g in RES respectively. 3445964k was free still. He was wondering why the system didn't use
2008 Feb 01
3
swapping on centos 5.1
Hi all, I used to use centos 4.5 on an AMD 4800+ with 2GIG ram. Now I use centos 5.1 on AMD 6400+ with 4GIG RAM. The system responsiveness is different between the two. I noticed that centos 5.1 seems to be swapping programs out of memory at times resulting in slowness (perceived by me). I played with swappiness (/proc/sys/vm/) setting to 10, then 1 then 0. Still resulted in the same perceived
2006 Jul 21
3
why is this machine using swap space?
Mem: 2075040k total, 1331452k used, 743588k free, 50896k buffers Swap: 2031608k total, 144k used, 2031464k free, 1191532k cached SunFire X2100 running Centos 4.3. Linux maytag1.texoma.net 2.6.9-34.0.2.EL #1 Fri Jul 7 19:24:57 CDT 2006 i686 athlon i386 GNU/Linux rgds/ldv
2012 Nov 15
3
Likely mem leak in 3.7
Starting with 3.7 rc1, my workstation seems to loose ram. Up until (and including) 3.6, used-(buffers+cached) was roughly the same as sum(rss) (taking shared into account). Now there is an approx 6G gap. When the box first starts, it is clearly less swappy than with <= 3.6; I can''t tell whether that is related. The reduced swappiness persists. It seems to get worse when I update
2012 May 08
2
How to deal with a dataframe within a dataframe?
Hello all, I am doing an aggregation where the aggregating function returns not a single numeric value but a vector of two elements using return(c(val1, val2)). I don't know how to access the individual columns of that vector in the resulting dataframe though. How is this done correctly? Thanks, robert > agg <- aggregate(formula=df$value ~ df$quarter + df$tool, + FUN=cp.cpk,
2018 Feb 05
0
Very slow rsync to gluster volume UNLESS `ls` or `find` scan dir on gluster volume first
Hi all, I have seen this issue as well, on Gluster 3.12.1. (3 bricks per box, 2 boxes, distributed-replicate) My testing shows the same thing -- running a find on a directory dramatically increases lstat performance. To add another clue, the performance degrades again after issuing a call to reset the system's cache of dentries and inodes: # sync; echo 2 > /proc/sys/vm/drop_caches I
2013 Aug 22
13
Lustre buffer cache causes large system overhead.
We have just discovered that a large buffer cache generated from traversing a lustre file system will cause a significant system overhead for applications with high memory demands. We have seen a 50% slowdown or worse for applications. Even High Performance Linpack, that have no file IO whatsoever is affected. The only remedy seems to be to empty the buffer cache from memory by running
2008 Jan 30
6
rsync and swapping
hi all, I use rsync to copy/backup ALL my stuff to another disk. When I run this seems like my machine (4 GIG ram centos 5.1) now begins to swap out more programs. Is there a way to reduce that swapping? I am running with echo 1 > /proc/sys/vm/swappiness I simply mount /dev/sdc1 /mnt/backup; mkdir /mnt/backup/month.day.year then rsync -a /home /mnt/backup/mon.day.year This is approximately
2015 Feb 24
2
Current 6.8 git build issues on HP-UX
On Wed, Feb 25, 2015 at 09:20:01AM +1100, Darren Tucker wrote: [...] > Sigh. And now the right patch from the tree that compiled. > (djm: I get the idea :-) Tim: is this sufficient to back out the "Work around finicky USL linker" change? https://anongit.mindrot.org/openssh.git/commit/?id=d1db656021d0cd8c001a6692f772f1de29b67c8b > diff --git a/openbsd-compat/bsd-misc.c
2004 Aug 06
0
admin 'rehash' broken in icecast 1.3.12?
I tried admin rehash for the first time since upgrading to 1.3.12 today. The rehash took at least 30 seconds [If not longer -- I didn't time it very closely] and icecast slowly snuck up to 99% CPU usage. I tried issuing the 'alias' command just to try and get a response from the console and that never responded. [I waited about two minutes] I had the same experience both of my
2004 Aug 06
0
admin 'rehash' broken in icecast 1.3.12?
Beau D Simensen <simensen@halogen.org> writes: > I tried admin rehash for the first time since upgrading to 1.3.12 today. The > rehash took at least 30 seconds [If not longer -- I didn't time it very > closely] and icecast slowly snuck up to 99% CPU usage. I tried issuing the > 'alias' command just to try and get a response from the console and that > never
2020 Mar 28
0
[klibc:update-dash] dash: exec: Never rehash regular built-ins
Commit-ID: e8b1ed5d253922aeb880518f27312b112e360f09 Gitweb: http://git.kernel.org/?p=libs/klibc/klibc.git;a=commit;h=e8b1ed5d253922aeb880518f27312b112e360f09 Author: Herbert Xu <herbert at gondor.apana.org.au> AuthorDate: Sat, 19 May 2018 02:39:51 +0800 Committer: Ben Hutchings <ben at decadent.org.uk> CommitDate: Sat, 28 Mar 2020 21:42:55 +0000 [klibc] dash: exec: Never
2003 Jun 04
1
Maybe a Rehash Call Queues
Hi All, I'm using the Call Queue without the Agent Login to provide a company with a Call Queue for their tech support Staff. Basically they have several techs the work from home. I dump the calls back out the PSTN via a SIP gateway. With the Queue application is there a way to define an absolute time out so that if a caller sits in the queue for say 10 minutes with out being serviced they
2007 Jun 23
0
[1060] trunk/wxruby2/swig/fixevents.rb: Include some previously missing events, rehash and simplify
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.1//EN" "http://www.w3.org/TR/xhtml11/DTD/xhtml11.dtd"> <html xmlns="http://www.w3.org/1999/xhtml"> <head><meta http-equiv="content-type" content="text/html; charset=utf-8" /><style type="text/css"><!-- #msg dl { border: 1px #006 solid; background: #369; padding:
2015 Jun 04
4
Effectiveness of CentOS vm.swappiness
Hi all, This might not be CentOS related at all. Sorry about that. I have lots of C6 & C7 machines in use and all of them have the default swappiness of 60. The problem now is that a lot of those machines do swap although there is no memory pressure. I'm now thinking about lowering swappiness to 1. But I'd still like to find out why this happens. The only common thing between all
2010 Mar 16
2
What kernel params to use with KVM hosts??
Hi all, I order to reach maximum performance on my centos kvm hosts I have use these params: - On /etc/grub.conf: kernel /vmlinuz-2.6.18-164.11.1.el5 ro root=LABEL=/ elevator=deadline quiet - On sysctl.conf # Special network params net.core.rmem_default = 8388608 net.core.wmem_default = 8388608 net.core.rmem_max = 16777216 net.core.wmem_max = 16777216
2015 Jun 05
2
Effectiveness of CentOS vm.swappiness
Am 05.06.2015 um 18:33 schrieb Gordon Messmer: > On 06/05/2015 03:29 AM, Markus "Shorty" Uckelmann wrote: >> some (probably unused) parts are swapped out. But, some of >> those parts are the salt-minion, php-fpm or mysqld. All services which >> are important for us and which suffer badly from being swapped out. > > Those two things can't really both be