On Fri, Dec 14, 2012 at 11:00 AM, scukaoyan@gmail.com
<scukaoyan@gmail.com>wrote:
> hi, all
>
> I have a master for about 3000 clients, runinternal is 20 min.
> A problem is that puppetdb''s mq is too long so that data in
postgre is the
> OLD version. any suggests to improve performance ?
>
We''ll need some more information in order to get a complete picture of
what''s going on. Can you send a screenshot of your puppetdb web console
after you''ve had it up for a few minutes? In particular, we''d
need to see
the rate at which your queue is growing, metrics around command processing
time and enqueue time, as well as information about catalog & resource
duplication rates. Also, what version of Postgres are you using, and have
you tuned any of the settings? Lastly, is the puppetdb system experiencing
a lot of iowait, or is it otherwised blocked on i/o or system time?
With that many nodes and a 20 minute runinterval, you''ll be sending
~2.5
catalogs every second to puppetdb...it''s not insurmountable, but
we''ll need
to look at the whole system in totality to see where the bottleneck is. :)
The first-order thing I''d look at is your catalog duplication rate...if
that''s low, then puppetdb is doing a lot of reads/writes to the
database
for every catalog received. Improving that number will help let puppetdb
short-circuit writing to disk at all, which should improve throughput
considerably.
You can also come and find us on IRC at #puppet on Freenode; that may make
for a faster back-and-forth.
Thanks!
--
deepak / Puppet Labs
--
You received this message because you are subscribed to the Google Groups
"Puppet Users" group.
To post to this group, send email to puppet-users@googlegroups.com.
To unsubscribe from this group, send email to
puppet-users+unsubscribe@googlegroups.com.
For more options, visit this group at
http://groups.google.com/group/puppet-users?hl=en.