Displaying 20 results from an estimated 40000 matches similar to: "mongrel garbage collection"
2007 Nov 05
29
Mongrel and memory usage
Hello,
I''m running a Rails application which must sort and manipulate a lot of data
which are loaded in memory.
The Rails app runs on 2 mongrel processes.
When I first load the app, both are 32Mb in memory.
After some days, both are between 200Mb and 300Mb.
My question is : is there some kind of garbage collector in Mongrel?
I never see the two Mongrel processes memory footprint
2006 Oct 31
9
Problems with mongrel dying
Hi
One of the two mongrel processes has died in the middle of the night
four times in the past 9 days, and I need help debugging this.
Each time the symptoms are the same:
* Each time I can restart the process via cap -a restart_app.
* Before the restart, there is nothing unusual in production.log or
mongrel.log.
* During the restart, about 100 repetitions of an error message are
generated in
2006 Sep 03
18
Recommentation: Sessions and PStore
Morning Folks,
As most of you know there were a few people who had the following three
bugs:
* CLOSE_WAIT: Lots of sockets in CLOSE_WAIT state.
* 99% CPU: Mongrel''s getting "stuck" pegged at 99% CPU.
* LEAK: Memory leak.
I''ve successfully fixed these bugs or attributed them to one main cause:
pstore.
First, the memory leak was because of a bug in how the GC in Ruby
2007 Oct 02
23
Mongrel using way more memory on production than staging. Any ideas why?
I''ve been trying to track down the culprit of erratic behaviour and
crashes on my production server (which is split into a number of Xen
instances), so set up a staging server so that I could really try to get
to the bottom of it.
The staging server (also split with Xen) is set up pretty much
identically as far as the mongrel_cluster server is concerned (the
production box has two
2006 Sep 04
2
The real reason why Sync and Mutex behave differently
As I''ve mentioned before, Sync and Mutex are very similar, and Mutex
is very simple.
Their locking algorithm (for exclusive locking) is essentially identical.
And in some detailed examinations of Mutex''s behavior, there''s nothing
superficially wrong with it. It''s pure ruby, so there are no funny
memory allocations at the C level, and it essentially operates
2008 Mar 12
12
Mongrel has crashed
Hi,
my mongrel has crashed with following errors in the logfile:
** Starting Rails with development environment...
** Rails loaded.
** Loading any Rails specific GemPlugins
** Signals ready. TERM => stop. USR2 => restart. INT => stop (no
restart).
** Rails signals registered. HUP => reload (without restart). It might
not work well.
** Mongrel 1.1.3 available at 0.0.0.0:3000
**
2007 Mar 06
59
Memory leaks in my site
Hi all,
My environment is ruby-1.8.4, rails 1.2.2, mongrel 1.0.1, linux 2.6. Now, i
have a problem on memory leaks with mongrel. My site is running 5 mongrel
processes on a 2G RAM machine, the memory of each process grows from about
20M to about 250M, but it never recover to the initial 20M, so i had to
restart the mongrel processes once per day. The load is about 1M hits per
day.
Waiting for
2006 Aug 25
9
Multi apps in one Mongrel instance
Hi,
It would be nice if Mongrel would do multiapps in one Mongrel instance.
For the moment, I have one Mongrel for each app of my server, each on a
different port.
For example:
http://server:8080/app1
http://server:8080/app2
instead of:
http://server:8080/
http://server:8081/
Is that planned?
Thanks.
-jec
--
JeSC - Software et Consulting
Jean-Eric Cuendet
Ing?nieur HES en
2007 Dec 14
21
Some silly benchs (was: 1.9)
Guys,
Just for fun, I tried to see (I know, a silly way to test it) how much
overhead we have calling the C functions of the extensions.
the benchmark script and the results:
http://pastie.caboo.se/128646
The naive C extension:
http://pastie.caboo.se/128647
I compared 1.8.6 (VC6 and mingw builds) against a fresh checkout of ruby trunk.
What I understand from that is 1.9 is slower than 1.8
2008 Mar 31
14
To surf to URL mydomain.com instead of mydomain.com:3000?
My production server is using mongrel_rails on Ubuntu linux.
With Firefox I can reach my production server with the URL:
www.mydomain.com:3000/
but I can''t reach my production server with the URL
www.mydomain.com/
Is there a way to configure mongrel so that I can reach my
production server with the URL www.mydomain.com?
--
Posted via http://www.ruby-forum.com/.
2007 May 21
13
swifty fly?
I heard about a new mongrel plugin or version that''s single threaded and
uses non-blocking IO with events. It''s called something like "swifty
fly". Anyone have a link for it?
Thanks.
2007 Dec 14
18
Mongrel error : EMFILE too many open files
I have a periodically_call_remote call in a partial running at 0.20
times a second.
<%= periodically_call_remote(:update => ''ack_distance_output'',
:frequency => 0.2,
:url=>{:action => :ackAdjustDistance})
%>
the action looks like this
def ackAdjustDistance
@calMessage = Calibration.getMessage
2006 Nov 14
16
memory issues -- mongrel part of the problem?
Hi. I''m a fairly new mongrel user and have a scenario that I wanted
to run by you all, see if anyone''s synapses connect -- mine def.
aren''t!
I''m running a fairly simple rails app through a mongrel cluster (2
mongrels). This site is consistently sloooow -- memory on the server
is regularly paging (and shouldn''t be). ssh-ing in and running top
shows
2006 Aug 07
4
Memory leak?
I am using BackgrounDRb to resize images after they are accepted into
our system. However, it has become clear that this creates a memory
leak. I''m not sure exactly where the leak exists, but I don''t think
it''s in my own code as it is presently being used in production and
has resized thousands of images without a leak occurring, thanks to
calling GC.start
2006 Jul 19
3
Recommended testing approaches for new Mongrel handlers?
Okay, I got off my duff this week and have been finishing Mongrel
support for IOWA. I now have two handlers that I need to test. One
is similar to the Rails handler, passing request handling to IOWA
threads in the same process, and the other allows one or more Mongrel
processes to pass requests to one or more separate IOWA processes,
with session affinity.
Is there any set of standard sorts of
2006 Sep 04
11
balancer://mongrel_cluster
A couple of questions please:
What is the benefit of having more than one Mongrel/Rails instance on
the same machine, could one instance not serve as many requests as say
three instances?
Is Mongrel Cluster a Apache add-in (like mod_*) or a separately running
load balancer which runs on a separate port, I could not work it out
from the sample Apache config...
|# Redirect all non-static
2007 Nov 01
8
Mongrel 1.1
Hello Mongrels,
Mongrel 1.1 is out with fullblown JRuby support, some reorganization,
and some bugfixes. Also, Mongrel_cluster has been updated to 1.0.4.
Go subscribe to the all-new news feed, because we probably won''t
announce to the list anymore:
http://mongrel.rubyforge.org/rss.xml
Thanks
Evan, and the Mongrel Team
--
Evan Weaver
Cloudburst, LLC
2006 Sep 01
8
Application performance
Hey all
Im working on improving the speed of my application. Ive taken on board
everything that has been said on this forum so far, im using httperf to
benchmark and try to improve performance. However, I have some question
marks over how best to go about improving performance in certain areas....
With no caching or such like deployed the application runs at an rather slow
35 - 38 req/s. A
2006 Sep 07
4
How to setup a sweeper to restart stale or hung mongrel servers
How do I setup a sweeper to restart stale or hung mongrel servers?
--
Jared Brown
jaredbrown at gmail.com
(765) 409-0875
7001 Central Ave
Indianapolis, IN 46220
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://rubyforge.org/pipermail/mongrel-users/attachments/20060907/0994017f/attachment.html
2007 May 22
4
Process safe log files with multiple mongrels
I have a rails application that runs on multiple mongrels and I would
like to make the log files non-interlacing. Are there any logging
modifications/other loggers out there that I could use to create
process safe logging?
Thank you,
Matt Margolis