Google should pay me to do a parody of this http://www.youtube.com/watch?feature=player_embedded&v=oz-7wJJ9HZ0#!
I might do it for free, but I suspect the audience for such a video is somewhat limited.
Dynamo Misses the Dynamic portion. Turn the dial to determine how many queries per second you need to serve from this non-relational Database. Got a Traffic Spike? Sucks to be you. Oh, they failed to mention, “changing the dial on a 100gig database?” it requires “about 1.5 minutes per gig when scaling up”. So those of you with 4TB of data, if you want to scale up you need to give them 4 days notice.
At the risk of someone tracking down my comments and using them against me… I talked to them the other day… They don’t get it.
My product has a version that runs on AWS and we tested it with Dynamo, it isn’t as fast as memcache. It’s not as fast as DataStore. They claimed they would be tuning it, but the it just isn’t as mature as Datastore.
A lot of why I started down the path of performance testing today was because Amazon CloudFront guys were talking about how great their product was and I said, it was slower than mine, and they came up with 2 countries (out of 12 tested) that were faster than me. But only by a few milliseconds. Switching to Route53, and now with a few optimization to how my app initializes I’m faster than they are again, (and you don’t need to understand anything about your infrastructure or your code to implement my solution).
Think about that. My software running on Google’s PaaS is faster and easier to implement than what I believe is the second largest CDN on the net. That is a powerful statement. Sure about me, but more about the power of the Google platform. Truly enterprise scale systems that perform like purpose built solutions. Amazon cost of delivery works out to be 16 cents a gig, we average 18. (we hope to have this to 16 with the new changes).
I don’t have the biggest deployment on AppEngine. I don’t spend the most money. There are for sure bigger and better stories out there, but Google is on the right track.
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To view this discussion on the web visit https://groups.google.com/d/msg/google-appengine/-/UbwFtEIY7scJ.
To post to this group, send email to google-a...@googlegroups.com.
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.
Their pricing is a confusing and means I have to estimate the capacity
I will need, which sucks. Most of the apps I deal with have pretty
distinct wave-like traffic patterns, the daily peaks see traffic
levels several times the levels during the troughs. I've also worked
with apps with a lot (several TB) of data that go from sustaining a
few hundred QPS to sustaining 1000+ QPS, fiddling with nobs while
that's happening would suck -- GAE just magically kept working (well
too!). I guess I just didn't get the dynamic part of DynamoDB.
However, I still hope that this will give Google incentive to innovate
and be competitive. In my opinion the datastore is/was probably App
Engines biggest competitive advantage. App Engine's datastore pricing
is also lacking, but at least your not paying for over capacity.
Robert
> --
> You received this message because you are subscribed to the Google Groups
> "Google App Engine" group.
You can't easily peer in to GAE's pricing. And a lot of people blame the
Expense on Google when they need to look at their own code. If we were
running the code from our first release we'd be paying 40x what we are now,
and we have maybe changed 60 lines of code. Some of those lines even slow
things down so we can offer more functionality, so likely if we had only
done feature expansion we'd be at 60x the price of what we are.
That's a big deal. It is also the kind of thing Google has never been good
at. They don't really do nice explanations of how to Optimize for their
products. I mean this in the most flattering of ways because the GAE team is
much more helpful than any other group I have interacted with at Google, but
generally Google's approach to documentation and evangelism is "We are so
much smarter than you we can't dumb this down enough that it would make
sense, and if you were smart enough to understand you wouldn't need me to
explain".
-----Original Message-----
From: google-a...@googlegroups.com
[mailto:google-a...@googlegroups.com] On Behalf Of stevep
Sent: Thursday, January 19, 2012 1:47 PM
To: Google App Engine
Subject: [google-appengine] Re: ROFLMFAO DynamoDB From Amazon
--
Have I Not shared when Google bot crawls me at 100k per hour? I hit 1/4 that
during those spikes. I could do some load testing if someone wants to foot
the bill, but I'll bet money I can push GAE to 1M easily.
-----Original Message-----
From: google-a...@googlegroups.com
[mailto:google-a...@googlegroups.com] On Behalf Of Andrei
Sent: Thursday, January 19, 2012 4:31 PM
To: Google App Engine
Subject: [google-appengine] Re: ROFLMFAO DynamoDB From Amazon
--
Or it may be the way your code is written, or the number of instances you
had active. What was the error you got when you pushed harder?
One of the problems we have been trying to work out with the CSV importer we
have been working on is that when we try to do 6 million records from our
command line python app, we are spinning up too many instances, so it costs
a fortune, but we have definitely been higher than 1500/s
Oh, they failed to mention, “changing the dial on a 100gig database?” it requires “about 1.5 minutes per gig when scaling up”. So those of you with 4TB of data, if you want to scale up you need to give them 4 days notice.
The overall time is not linear as the Google Groups poster suggests. In most cases it will be between a few minutes to a few hours regardless of total size. Larger data sets may take a bit longer than smaller data sets simply because there is often more data movement to perform and coordination to be made across a greater number of machines. Rest assured though, we make use of parallelism where we can so the curve is far from linear.
I was talking to them. They do On demand replication based on scale. The formula is a bit more complex than what I laid out, but if you needed to scale up for a “slash dot” you couldn’t.
From: google-a...@googlegroups.com [mailto:google-a...@googlegroups.com] On Behalf Of a its_me
Sent: Friday, January 20, 2012 3:19 PM
To: google-a...@googlegroups.com
Subject: [google-appengine] Re: ROFLMFAO DynamoDB From Amazon
Oh, they failed to mention, “changing the dial on a 100gig database?” it requires “about 1.5 minutes per gig when scaling up”. So those of you with 4TB of data, if you want to scale up you need to give them 4 days notice.
I've been searching... where did you find that? (Link?)
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To view this discussion on the web visit https://groups.google.com/d/msg/google-appengine/-/Unne6VNwYxkJ.
We write any request that has been expired (file hasn’t been accessed in X seconds) or any new request we haven’t seen.
While we average for the most part about 99% of requests from Users being served without touching the data store, when Google bot indexes a site it is often hitting pages users never do.
JeffProbst.com has 35 total pages, 500 total assets. We don’t have to touch the datastore (or memcache) for 99.998% of requests
XYHD.tv has 4800 total pages 8400 total assets. In a 24 hour period 350 unique pages receive user traffic 800-ish assets. When Google Bot comes through it reads 4200 ish of those pages. On XYHD.tv pages expire from the cache every 3 minutes. So we get about 97% cache hits for users, but only 10% for Google bot.
From: google-a...@googlegroups.com [mailto:google-a...@googlegroups.com] On Behalf Of Niklas Rosencrantz
Sent: Saturday, January 21, 2012 3:45 PM
To: google-a...@googlegroups.com
Subject: [google-appengine] Re: ROFLMFAO DynamoDB From Amazon
Thanks for this thread. I'm curious why you need so many writes. Do you write to the datastore just because there is a request? Isn't that ineffeicient? I have many handlers that don't do writes at all.
Best regards,
Nick
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To view this discussion on the web visit https://groups.google.com/d/msg/google-appengine/-/XkvAJNYnYUQJ.
Jeff
Those numbers are approximate, but close.
Jeff
--
You received this message because you are subscribed to the Google Groups
"Google App Engine" group.
-----Original Message-----
From: google-a...@googlegroups.com
[mailto:google-a...@googlegroups.com] On Behalf Of Jeff Schnitzer
Sent: Monday, January 23, 2012 12:56 PM
To: google-a...@googlegroups.com
Jeff
--
You received this message because you are subscribed to the Google Groups
"Google App Engine" group.