similar to: murdering high-memory workers and auto-scaling

Displaying 20 results from an estimated 4000 matches similar to: "murdering high-memory workers and auto-scaling"

2010 Oct 29
2
stop/start workers "by need"
Hello to everybody. I making shared hosting for rails and rack applications. For app. server i userd phusion passenger, but after some issues with that, want to migrate to unicorn. Is there any way to make unicorn stop all workers and start them on first request? It''ll be really great to have this opportunity. >No subscription is needed to post to the mailing list, let us know
2012 Nov 29
13
Fwd: Maintaining capacity during deploys
We''re using unicornctl restart with the default before/after hook behavior, which is to reap old Unicorn workers via SIGQUIT after the new one has finished booting. Unfortunately, while the new workers are forking and begin processing requests, we''re still seeing significant spikes in our haproxy request queue. It seems as if after we restart, the unwarmed workers get swamped by
2011 May 19
2
unicorn doesn't restart properly after cap deploy (not using Bundler)
Hey guys, I''m sending USR2 to unicorn after cap deploy, and the old master is getting replaced by a new master that doesn''t work. Here are some similar threads I found ... http://rubyforge.org/pipermail/mongrel-unicorn/2010-October/000733.html http://rubyforge.org/pipermail/mongrel-unicorn/2010-October/000717.html I''m not using Bundler, and I have working_directory
2012 Jun 27
3
Unicorn workers under Monit
Hi there, I would like to config Monit to monitor our production-unicorn-workers What memory size would you recommend to be the maximum reachable for a worker? so Monit can restart it. Thanks a lot ! Manuel Palenciano
2012 Apr 12
8
Background jobs with #fork
Hi I''ve migrated from Passenger to Unicorn about a week ago. It''s great. Great transparency and management, thanks for this great software! A few of my Rails applications start background jobs using Kernel#fork. Of course, the ActiveRecord connections are closed and reopened again in the parent and child processes. The child process also does its job. Unfortunately, it seems
2009 Oct 13
7
Unicorn Nginx Issue
I''ve setup nginx with this server { listen 80; server_name unicorn.local; location / { proxy_pass http://unicorn; } } upstream unicorn { server unix:/Users/mattmongeau/projects/test/unicorn/tmp/sockets/unicorn.sock; } When I run unicorn_rails -c config/unicorn.rb -E development I can access the application just fine on unicorn.local When I run unicorn_rails -c
2011 Dec 08
5
Master repeatedly killing workers due to timeouts
Hi, We''re using unicorn as a Rails server on Solaris, and it''s been running great for several months. We''ve recently been having a few problems and I''m at a loss what might cause it. A number of times in the past few days, our unicorn slaves keep timing out & the master keeps restarting them. unicorn.log looks something like : E,
2009 Feb 11
16
Unicorn: UNIX+localhost/LAN-only Mongrel fork
Hello all, Last week, I finally decided to put into motion some ideas I''ve been kicking around for a year in my head since last year... Basically I don''t want to have to deal with threads or support platforms that rely on or encourage threads. Especially given MRI 1.9 where kernel threads are more difficult to debug than green ones. Given the limited scope of this project,
2013 Dec 09
2
[PATCH] rework master-to-worker signaling to use a pipe
Signaling using normal kill(2) is preserved, but the master now prefers to signal workers using a pipe rather than kill(2). Non-graceful signals (:TERM/:KILL) are still sent using kill(2), as they ask for immediate shutdown. This change is necessary to avoid triggering the ubf (unblocking function) for rb_thread_call_without_gvl (and similar) functions extensions. Most notably, this fixes
2013 May 20
2
Unicorn + RUnit Rails Not Killing Old Master
Hi, I''m deploying Unicorn on a Rails application with RUnit. Technically I''m using Chef''s deployment tools, if any of you are familiar with it (https://github.com/opscode-cookbooks/application_ruby) but to be clear they aren''t doing anything magical, so this is purely an issue with RUnit and Unicorn. The TL;DR of the following post, which has lots of
2011 Sep 30
3
Multiple rack applications on the same server with unicorn
If I''m running two rails apps on the same server using Unicorn, I have to run two instances of Unicorn, right? If so, then here''s a place where passenger might win in terms of memory use, as the rails code will be loaded into memory twice, right? I''m still probably going with Unicorn, but just exploring this first. Thanks, John
2011 Sep 06
1
Workers getting broken after some time
I have strange error that happens after couple of hours. Some instances start to respond on any request with status code 500. Log file unicorn.stderr.log gets populated with this error: app error: undefined method `each'' for nil:NilClass (NoMethodError) /usr/local/lib/ruby/gems/1.9.1/gems/rack-1.1.2/lib/rack/utils.rb:267:in `initialize''
2011 Jan 13
5
Unicorn 3.3.1 "Too many open files" error in kgio_tryaccept
Strange error which took down my server today: Unhandled listen loop exception #<Errno::EMFILE: Too many open files - accept>. /usr/local/ruby-enterprise-1.8.7-2010.01/lib/ruby/gems/1.8/gems/unicorn-3.3.1/lib/unicorn/http_server.rb:606:in `kgio_tryaccept'' /usr/local/ruby-enterprise-1.8.7-2010.01/lib/ruby/gems/1.8/gems/unicorn-3.3.1/lib/unicorn/http_server.rb:606:in
2011 Aug 16
6
Unicorn logging in production env
Hi All, as I can see on https://github.com/defunkt/unicorn/blob/master/lib/unicorn.rb#L53 the?Rack::CommonLogger used only in development env and in weird evn called "deployment". Any chance to add "production" to this case? Serg Podtynnyi
2013 Aug 20
14
A barrage of unexplained timeouts
We''ve been running unicorn-3.6.2 on REE 1.8.7 2011.12 in production for quite some time and we use monit to monitor each unicorn worker. Occasionally, I''ll get a notification that a worker has timed-out and has been re-spawned. In all these cases, when I look at the rails logs, I can see the last request that the worker handled, and they all have appeared to complete
2013 Feb 07
5
does unicorn fork like phusion?
Does unicorn work like phusion in that it automatically forks based on traffic levels? Or is that what thin does also? -- You received this message because you are subscribed to the Google Groups "Ruby on Rails: Talk" group. To unsubscribe from this group and stop receiving emails from it, send an email to rubyonrails-talk+unsubscribe-/JYPxA39Uh5TLH3MbocFF+G/Ez6ZCGd0@public.gmane.org
2012 Nov 07
1
select(): Interrupted system call from curb when stopping unicorn
Hi, We''ve just migrated one of our rails applications from nginx/passenger running on REE 1.8.7 Ubuntu 8.04 to unicorn running on MRI 1.9.3 on Ubuntu 10.04. The app makes a number of calls to internal services using curb. Our deployment script stops unicorn by sending SIGQUIT to the unicorn master, sleeps for a few seconds to ensure that HAProxy has taken the node out of service and
2010 Sep 09
4
Unicorn fails to restart gracefully on capistrano deploy
First off thanks very much for all the hard work on unicorn. Alas, we''ve encountered an issue where unicorn fails to spawn new workers that have loaded the incoming revision on a capistrano deploy. I''m not entirely sure the issue is due to unicorn as it appears that bundler was responsible for a similar issue in the past:
2012 Oct 18
4
Rack env rack.multiprocess true with single worker
Hi, unicorn unconditionally sets rack.multiprocess to true in the Rack environment. The Rack spec [0] says the following about the variable: "true if an equivalent application object may be simultaneously invoked by another process, false otherwise." When unicorn is running with a single worker this does not hold so what do you think about setting the variable to false when only a
2010 Jul 13
4
SIGWINCH
Hi, I followed the procedure to replace a running unicorn. Works fine, except for one thing: after I decide to back out, ie I send a HUP to the old master followed by a QUIT to the new master, then: 1. the new master does die, and the old master does work again 2. the "old" master keeps running with a process name of "master (old)" 3. the "old" master keeps