What sorts of steps have you already taken in order to investigate the
problem? Does the Puppetmaster have enough RAM? It is otherwise
overloaded from a system load perspective? What about those two factors
on the client(s)? In other words, have you eliminated all of the
standard system-based considerations?
If so, I would recommend running the client manually with test and debug
enabled, and taking a look at the output there. As well, if it's
possible, you may wish to do the same for the master daemon - that will
give you a much better idea of where along the process any potential
hangs are occurring.
Good luck !
--
Daniel Maher
� makin' plans now to live on Mars 'cuz I got Earth on lock. �
The obvious suggestion is to upgrade :-).
ken.
> --
> You received this message because you are subscribed to the Google Groups "Puppet Users" group.
> To post to this group, send email to puppet...@googlegroups.com.
> To unsubscribe from this group, send email to puppet-users...@googlegroups.com.
> For more options, visit this group at http://groups.google.com/group/puppet-users?hl=en.
>
>
I guess I was looking to find out the specifics of how big these
manifests and templates are in total on the nodes you are having
specific performance issues with. You say 'big manifests' but this is
a relative statement :-).
If 'nothing has changed' (please confirm this with your svn/git
history) then it must be something else other then content or puppet
itself ....
Without seeing your content its a hard one to trace via a mailing
list, there could be a network issue and some function you are trying
to execute could be lagging or having trouble resolving dns etc. Try
running your puppetmasterd in the foreground using something like:
puppetmasterd --no-daemonize --debug --trace --logdest console
And as you do a single run, see if it gives you any clues as to what
its blocking on during those 20 minutes.
ken.
ken.
I'm pretty sure those entries you are looking at are file resources. In general, file resources scale very poorly with webrick. Even without webrick, they still tend to perform poorly.
My advice:
Stop using webrick
Take a look at recuse=>remote if you use recursive file copies a lot
Make sure you're not swapping and that you have enough ram
Consider copying big files or directories with a lot of files with a package or tar+wget.
> as i m able to see that puppet master taking around 7-8 seconds to
> compiler catalog
> and puppet client taking around 1106 second to apply catalog .
run the client with --summarize and see on which resource types it is
spending most of the time, then you can probably figure out what is wrong.
also analyzing the reports with something like the dashboard or the
foreman can help to trace these issues.
Certainly something within your environment or its dependencies has
changed. Puppet is heavily calling out to tools like facter, package
managers, service scripts etc. which all can easily affect your puppet runs.
~pete
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.11 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/
iEYEARECAAYFAk4R+mcACgkQbwltcAfKi3/d/QCeODSqHvF8OxULMcbw6gfuEOT9
noMAn2b0DySrqrZ27kA6Xfvb1WaVWKdX
=pHX/
-----END PGP SIGNATURE-----
Any network changes?
Do you have a switch plugged into itself?
what does tcpdump show? Normal activity?
--
vagn
On 07/05/2011 12:11 AM, sanjiv.singh wrote:
> some times puppet client fails with following.......
>
> debug: Loaded state in 0.01 seconds
> debug: Using cached certificate for ca
> debug: Using cached certificate for pb11-9.domain.local
> debug: Using cached certificate_revocation_list for ca
> debug: catalog supports formats: b64_zlib_yaml marshal pson raw yaml;
> using pson
> err: Could not retrieve catalog from remote server: Could not intern
> from pson: expected value in object at '"File[/var/lib/'!
> warning: Not using cache on failed catalog
> err: Could not retrieve catalog; skipping run
>
> thanks,
> Sanjiv Singh
>
> File: 1539.03
This indicates clearly a problem in serving files, so does the other
error you posted.
As others asked: Are you still running with Webrick (means default
puppetmaster without setting up passenger/mongrel)? And how many
clients are connecting to the server how frequently?
~pete
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.11 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/
iEYEARECAAYFAk4SpsUACgkQbwltcAfKi3+AeQCfVKw/DBKcMIlCzMfGK7gF0xpK
0BYAn3ph1iWV8YjO5WxuEzm7bhucvAU+
=Aj9t
-----END PGP SIGNATURE-----
One question - does this happen each time you run it on the same box?
If there is no new changes - is the performance still terrible or is
it only for systems where there is a new change?
I'm just pondering also - did you introduce a new large file? I'm just
wondering if this time is spent doing an md5 sum of a huge file for
file {} comparison. You see the file provider will do an md5
comparison to make sure a a file that already exists doesn't get
installed again. If the file is large - then it takes a long time.
Does this slowness happen even when it appears no changes need to be
made? If you think a file is a culprit you can emulate the md5
behaviour by running 'time md5sum bigassfile.iso' to see if it takes a
long time.
Otherwise its the delivery of such a large file ... which might be the
case, and you definitely should switch out your stack.
I see you have 211 resources or so ... so its not the quantity of
files here that is the problem I would guess.
> There are around 100+ servers in my environment.but at most puppet
> master serving to 10-12 nodes (varies) concurrently.
Webrick at best can manage 1 node concurrently properly. So the minute
you fire up more then 1 it will slow down as it tries to serve each
client when available. Switch to Apache/Mongrel :-).
ken.