I have seen the same thing… Not to the same degree. Other than saying it happens… I don’t know what to tell you. Report it, and Google will fix it, they usually detect and fix it on their own. Only thing that annoys me is that my CPU usage is more expensive during these times so I’m slow AND I am paying more for my service.

Brandon Wirtz |
| |||
| ||||
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To post to this group, send email to google-a...@googlegroups.com.
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.
Like Brandon said, file a production issue and they'll get on it.
http://code.google.com/p/googleappengine/issues/entry?template=Production%20issue
Robert
I asked about other services because sometimes they will experience
elevated latency. Depending on what you're using and how your code is
structured, small increases in latency could be significant.
Robert
What I've found is that unless you've got very spiky traffic, that
suddenly and massively spikes, you can turn the max idle instances
down. They more-or-less just act like a buffer, but don't actively
serve traffic. If you want better performance turn the min-idle
instances up a little, but again this doesn't need to be too high. I
generally run apps with between "automatic" and 15 idle instances.
Experiment to find the right number.
To get better performance I generally leave the min pending latency on
automatic, but I reduce the max pending latency. For reasonable apps
a setting somewhere between 300ms and 600ms seems to do a good job,
but you'll be spinning up instances for more rapidly. If you don't
want to, bump it up to a max of 1000ms (and possibly increase the min
to a couple hundred ms).
These are just my observations from experimenting. Your results might
be different.
Robert
I find that you should very, very rarely need anywhere near that.
I did a post a while back with some rough formulas, but if your Max idle
time is 30 seconds minus (6x your average request time + warmup time), you
can generally weather any spike.
Typically if you set max pending to 1s your users will be a lot happier, and
you don't have to do as much with tuning. I used to spend more time tuning
max idle instances, now, I set that to 1 or 2 (depending on my instance
size) and tune the max idle time. (lower means better scale)
Did you factor into account app startup times? Some apps can easily
take several seconds to get fired up, so they do better with a larger
buffer. For light apps I agree, I've found 1 or 2 is sufficient.
>
> Typically if you set max pending to 1s your users will be a lot happier, and
> you don't have to do as much with tuning. I used to spend more time tuning
> max idle instances, now, I set that to 1 or 2 (depending on my instance
> size) and tune the max idle time. (lower means better scale)
I agree here too, I generally set my max idle to 5 then don't mess
with it. I prefer 500ms for my default max latency, but a lot of my
stuff is tuned to spin up fast.
Yeah, I said warm-up not startup, but same thing.
I like lower numbers too, but if your max latency is at 1/3 your average you
will pretty much be guaranteed to have an instance per simultaneous request.
My average request is 300ish MS so I can pull off 500-600ms, but someone
with 2.5s average would spin up a lot of extra instances everytime a user
hit a page making 8 requests. And you'd eat the start up time on all those
instances, so too low of a Max can make things worse. Again check the
archive I talked about this too.