Here's the line from my dashboard:
Total Stored Data $0.005/GByte-day 82% 7.42 of 9.00 GBytes
$0.04 / $0.04
And here is the word from my datastore statistics:
Last updated Total number of entities Size of all entities
1:32:13 ago 232,867 615 MBytes
(metadata 11%, if that matters)
Please, can someone help me figure out this issue? I'd be happy to
share any info or code which would help track this down. My app id is
vulahealth.
I can not find the post that referenced this though.
On Mon, Mar 22, 2010 at 12:21 PM, Tom Wu <servic...@gmail.com> wrote:
> GAE is cluster which include master & slaver, backup system... etc..
> So the quota is much bigger than your local file.
>
>
>
> 2010/3/22 杨浩 <skzr...@gmail.com>
>>
>> I have the same about thaTt!
>> --
>> You received this message because you are subscribed to the Google Groups
>> "Google App Engine" group.
>> To post to this group, send email to google-a...@googlegroups.com.
>> To unsubscribe from this group, send email to
>> google-appengi...@googlegroups.com.
>> For more options, visit this group at
>> http://groups.google.com/group/google-appengine?hl=en.
>
> --
> You received this message because you are subscribed to the Google Groups
> "Google App Engine" group.
> To post to this group, send email to google-a...@googlegroups.com.
> To unsubscribe from this group, send email to
> google-appengi...@googlegroups.com.
> For more options, visit this group at
> http://groups.google.com/group/google-appengine?hl=en.
>
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To post to this group, send email to google-a...@googlegroups.com.
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.
Clearly, I can fix that moving forward, though it will cost me a lot
of CPU to fix the data I've already entered. But as a short-term
stopgap, is there any way to delete entire default indexes for a given
property? (I mean, anything besides setting indexed=False and then
touching each entity one-by-one). You can vacuum custom indexes - can
you do it with indexes created by default?
Thanks,
Jameson
On 22 mar, 03:42, "Nick Johnson (Google)" <nick.john...@google.com>
wrote:
> > google-appengi...@googlegroups.com<google-appengine%2Bunsubscrib e...@googlegroups.com>
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
1. Is the overhead for erasing data (and thus whittling down indexes)
over half the overhead from adding it? Under 10%? Or what? (I don't
need exact numbers, just approximates.
2. If it's more like half - is there some way to just nuke all my data
and start over?
Thanks,
Jameson
On 22 mar, 03:42, "Nick Johnson (Google)" <nick.john...@google.com>
wrote:
> > google-appengi...@googlegroups.com<google-appengine%2Bunsubscrib e...@googlegroups.com>
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
OK, after hashing it out on IRC, I see that I have to erase my data
and start again.
Since it took me 3 days of CPU quota to add the data,
I want to know if I can erase it quickly.
1. Is the overhead for erasing data (and thus whittling down indexes)
over half the overhead from adding it? Under 10%? Or what? (I don't
need exact numbers, just approximates.
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
On Mar 22, 3:48 pm, "Nick Johnson (Google)" <nick.john...@google.com>
wrote
> On Mon, Mar 22, 2010 at 8:45 PM, homunq <jameson.qu...@gmail.com> wrote:
> > OK, after hashing it out on IRC, I see that I have to erase my data
> > and start again.
>
> Why is that? Wouldn't updating the data be a better option?
Because everything about it is wrong for saving space - the key names,
the field names, the indexes, and even in one case the fact of
breaking a string out into a list. (something I did for better
searching in several cases, one of which is not worth it now I realize
that 10X is easy to hit.)
And because the data import runs smoothly, and I have code for that
already.
....
Watching my deletion process start to get trapped in molasses, as Eli
Jones mentions above, I have to ask two things again:
1. Is there ANY ANY way to delete all indexes on a given property
name? Without worrying about keeping indexes in order when I'm just
paring them down to 0, I'd just be running through key names and
deleting them. It seems that would be much faster. (If it's any help,
I strongly suspect that most of my key names are globally unique
across all of Google).
2. What is the reason for the slowdown? If I understand his suggestion
to delete every 10th record, Eli Jones seems to suspect that it's
because there's some kind of resource conflict on specific sections of
storage, thus the solution is to attempt to spread your load across
machines. I don't see why that would cause a gradual slowdown. My best
theory is that write-then-delete leaves the index somehow a little
messier (for instance, maybe the index doesn't fully recover the
unused space because it expects you to fill it again) and that when
you do it on a massive scale you get massively messy and slow indexes.
Thus, again, I suspect this question reduces to question 1, although I
guess that if my theory is right a compress/garbage-collect/degunking
call for the indexes would be (for me) second best after a way to nuke
them.
On Mar 22, 3:48 pm, "Nick Johnson (Google)" <nick.john...@google.com>
wrote
> On Mon, Mar 22, 2010 at 8:45 PM, homunq <jameson.qu...@gmail.com> wrote:Because everything about it is wrong for saving space - the key names,
> > OK, after hashing it out on IRC, I see that I have to erase my data
> > and start again.
>
> Why is that? Wouldn't updating the data be a better option?
the field names, the indexes, and even in one case the fact of
breaking a string out into a list. (something I did for better
searching in several cases, one of which is not worth it now I realize
that 10X is easy to hit.)
And because the data import runs smoothly, and I have code for that
already.
....
Watching my deletion process start to get trapped in molasses, as Eli
Jones mentions above, I have to ask two things again:
1. Is there ANY ANY way to delete all indexes on a given property
name? Without worrying about keeping indexes in order when I'm just
paring them down to 0, I'd just be running through key names and
deleting them. It seems that would be much faster. (If it's any help,
I strongly suspect that most of my key names are globally unique
across all of Google).
2. What is the reason for the slowdown? If I understand his suggestion
to delete every 10th record, Eli Jones seems to suspect that it's
because there's some kind of resource conflict on specific sections of
storage, thus the solution is to attempt to spread your load across
machines. I don't see why that would cause a gradual slowdown. My best
theory is that write-then-delete leaves the index somehow a little
messier (for instance, maybe the index doesn't fully recover the
unused space because it expects you to fill it again) and that when
you do it on a massive scale you get massively messy and slow indexes.
Thus, again, I suspect this question reduces to question 1, although I
guess that if my theory is right a compress/garbage-collect/degunking
call for the indexes would be (for me) second best after a way to nuke
them.
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To post to this group, send email to google-a...@googlegroups.com.
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.
It seems to me that having no index at all is the same situation as if
the property was indexed=False from the beginning. If that's so, it
can't be violating a hard constraint.
>
> > 2. What is the reason for the slowdown? If I understand his suggestion
> > to delete every 10th record, Eli Jones seems to suspect that it's
> > because there's some kind of resource conflict on specific sections of
> > storage, thus the solution is to attempt to spread your load across
> > machines. I don't see why that would cause a gradual slowdown. My best
> > theory is that write-then-delete leaves the index somehow a little
> > messier (for instance, maybe the index doesn't fully recover the
> > unused space because it expects you to fill it again) and that when
> > you do it on a massive scale you get massively messy and slow indexes.
> > Thus, again, I suspect this question reduces to question 1, although I
> > guess that if my theory is right a compress/garbage-collect/degunking
> > call for the indexes would be (for me) second best after a way to nuke
> > them.
>
> Deletes using the naive approach slow down because when a record is deleted
> in Bigtable, it simply inserts a 'tombstone' record indicating the original
> record is deleted - the record isn't actually removed entirely from the
> datastore until the tablet it's on does its next compaction cycle. Until
> then, every subsequent query has to skip over the tombstone records to find
> the live records.
>
> This is easy to avoid: Use cursors to delete records sequentially. That way,
> your queries won't be skipping the same tombstoned records over and over
> again - O(n) instead of O(n^2)!
>
Thanks for explaining. Can you say anything about how often the
compaction cycles are? Just an order of magnitude - hours, days, or
weeks?
Thanks,
Jameson
It seems to me that having no index at all is the same situation as if
>
> > Watching my deletion process start to get trapped in molasses, as Eli
> > Jones mentions above, I have to ask two things again:
>
> > 1. Is there ANY ANY way to delete all indexes on a given property
> > name? Without worrying about keeping indexes in order when I'm just
> > paring them down to 0, I'd just be running through key names and
> > deleting them. It seems that would be much faster. (If it's any help,
> > I strongly suspect that most of my key names are globally unique
> > across all of Google).
>
> No - that would violate the constant that indexes are always kept in sync
> with the data they refer to.
>
the property was indexed=False from the beginning. If that's so, it
can't be violating a hard constraint.
Thanks,
Jameson
--
You received this message because you are subscribed to the Google Groups "Google App Engine" group.
To post to this group, send email to google-a...@googlegroups.com.
To unsubscribe from this group, send email to google-appengi...@googlegroups.com.
For more options, visit this group at http://groups.google.com/group/google-appengine?hl=en.