As you might already have heard, we have a big announcement: Metaweb is joining Google!
You can read about it on our blog and on Google’s blog:
http://blog.freebase.com/2010/07/16/metaweb-joins-google/
http://googleblog.blogspot.com/2010/07/deeper-understanding-with-metaweb.html
As you’ll see from both blog posts, Google is committed to keeping Freebase free and open. In fact, one thing that’s happening is that effective immediately, we’re making our data dumps weekly instead of quarterly. As for everything else, right now things will continue pretty much as usual, and going forward we’re looking forward to making Freebase bigger and better with Google's help.
As you can imagine, this is a ridiculously busy time for us as we settle in at Google, so please be understanding with us for a little while. In the meantime, if you have any Freebase-related questions, please feel free to drop me an email (direct at kir...@metaweb.com if you like) or find me on IRC (irc.freenode.net #freebase).
K.
--
Kirrily Robert
Freebase Community Director
kir...@metaweb.com
Shawn
Kirrily Robert wrote:
> Hey everyone,
>
> As you might already have heard, we have a big announcement: Metaweb is joining Google!
>
> You can read about it on our blog and on Google�s blog:
> http://blog.freebase.com/2010/07/16/metaweb-joins-google/
> http://googleblog.blogspot.com/2010/07/deeper-understanding-with-metaweb.html
>
> As you�ll see from both blog posts, Google is committed to keeping Freebase free and open. In fact, one thing that�s happening is that effective immediately, we�re making our data dumps weekly instead of quarterly. As for everything else, right now things will continue pretty much as usual, and going forward we�re looking forward to making Freebase bigger and better with Google's help.
> Congratulations to everyone at Metaweb! This is fantastic news. Its
> great to see that all your hard work has paid off and that Freebase
> will continue to grow and improve under Google. This is certainly an
> exciting time for all of us to be part of the open data movement.
>
> Shawn
>
> Kirrily Robert wrote:
>> Hey everyone,
>>
>> As you might already have heard, we have a big announcement:
>> Metaweb is joining Google!
>>
>> You can read about it on our blog and on Google’s blog:
>> http://blog.freebase.com/2010/07/16/metaweb-joins-google/
>> http://googleblog.blogspot.com/2010/07/deeper-understanding-with-metaweb.html
>>
>> As you’ll see from both blog posts, Google is committed to keeping
>> Freebase free and open. In fact, one thing that’s happening is
>> that effective immediately, we’re making our data dumps weekly
>> instead of quarterly. As for everything else, right now things
>> will continue pretty much as usual, and going forward we’re looking
>> forward to making Freebase bigger and better with Google's help.
>>
>> As you can imagine, this is a ridiculously busy time for us as we
>> settle in at Google, so please be understanding with us for a
>> little while. In the meantime, if you have any Freebase-related
>> questions, please feel free to drop me an email (direct at kir...@metaweb.com
>> if you like) or find me on IRC (irc.freenode.net #freebase).
>>
>> K.
>>
>>
>
> _______________________________________________
> You are receiving this message because you are subscribed to the
> Freebase-discuss mailing list.
> To post a message to the list: Freebase...@freebase.com
> To unsubscribe, view archives, etc: http://lists.freebase.com/mailman/listinfo/freebase-discuss
Will transition to *really* excited about this as soon as the DOJ says it's OK for us to be colleagues!
glenn (Needle/ITA)
Excited about this for both Metaweb and Google. Will transition to *really* excited about this as soon as the DOJ says it's OK for us to be colleagues! glenn (Needle/ITA) On 16 Jul 10, at 2:18pm, Kirrily Robert wrote:
Hey everyone, As you might already have heard, we have a big announcement: Metaweb is joining Google! You can read about it on our blog and on Google�s blog: http://blog.freebase.com/2010/07/16/metaweb-joins-google/ http://googleblog.blogspot.com/2010/07/deeper-understanding-with-metaweb.html As you�ll see from both blog posts, Google is committed to keeping Freebase free and open. In fact, one thing that�s happening is that effective immediately, we�re making our data dumps weekly instead of quarterly. As for everything else, right now things will continue pretty much as usual, and going forward we�re looking forward to making Freebase bigger and better with Google's help. As you can imagine, this is a ridiculously busy time for us as we settle in at Google, so please be understanding with us for a little while. In the meantime, if you have any Freebase-related questions, please feel free to drop me an email (direct at kir...@metaweb.com if you like) or find me on IRC (irc.freenode.net #freebase). K. -- Kirrily Robert Freebase Community Director
kir...@metaweb.com _______________________________________________ You are receiving this message because you are subscribed to the Freebase-discuss mailing list. To post a message to the list: Freebase...@freebase.com To unsubscribe, view archives, etc: http://lists.freebase.com/mailman/listinfo/freebase-discuss
That is just what I thought! :) The global presence of Google is something that could push the I18N and I10N of Freebase at some future date.
-hangy
congrats to everyone,
I hope you'll be able to keep up your great attitude inside the new
corporate environment and get the most of it. The amount of stuff that
could be poured into FB from base/froogle/maps/knol/etc is mindbending
:)
> You can read about it on our blog and on Google’s blog:
> http://blog.freebase.com/2010/07/16/metaweb-joins-google/
> http://googleblog.blogspot.com/2010/07/deeper-understanding-with-metaweb.html
<snip>
I noticed the thing about the data dumps becoming weekly, that seems
quite a difference :)
But (I seem to recall this was discussed in the past but can't find
it) if they are this frequent it seems even more useful to provide
downloads of deltas mext to the big packages. Any hope? :)
The diff between two consecutive dumps, to avoid downloading a great big
file, of which 99% is exactly the same data as you downloaded a week
ago.
If we're going to weekly dumps, it may just about be feasible to
construct the deltas via MQL.
Cheers,
Phil
--
Philip Kendall <phi...@shadowmagic.org.uk>
http://www.shadowmagic.org.uk/
On Sat, Jul 17, 2010 at 08:44:29AM -0500, Thad Guidry wrote:Curious. What kind of deltas would be appealing to you ? Can you give examples ?The diff between two consecutive dumps, to avoid downloading a great big file, of which 99% is exactly the same data as you downloaded a week ago. If we're going to weekly dumps, it may just about be feasible to construct the deltas via MQL. Cheers, Phil
Exactly,
since the data dumps are in TSV format, they are suitable for line
oriented diffs, which can be individually downloaded incrementally
(and trivially scripted on our side if someone wants to fetch/apply
them less frequently).
For the quarterly dump it may have little sense because a lot of data
would be different, but for one week i guess most content would be
stable.
That's an intriguing idea. We internally use a notification stream for replicating changes across graphds and into the search index (for Suggest). I don't know technically how big a leap it would be to go from that to syndicating that stream externally, though. Worth looking into.
-jason
On Sat, Jul 17, 2010 at 8:09 PM, Jason Douglas <ja...@metaweb.com> wrote:
>> My personal preference would be PubSubHub. That way you just get the deltas you're interested in as they happen.
>>
>> Shawn
>
> That's an intriguing idea. We internally use a notification stream for replicating changes across graphds and into the search index (for Suggest). I don't know technically how big a leap it would be to go from that to syndicating that stream externally, though. Worth looking into.
I believe these are different issues:
1. being able to be always up to date for a msq-selected subset (I
second the PuSH suggestion), which would be ubercool
2. being able to to obtain the full data in an efficient and
incremental fashion, which is simpler but solves my actual problem :)
To be clear: in one of our uses of the FB data we need to process the
whole dump to select a large subset of the topics, integrate it with
our data and compute some metrics, which is a batch process.
Even if we had updates available every minute we would still be
performing it as a scheduled process so 1. would not be especially
useful because I'd need to implement "stashing" on my side, while 2.,
saving a large part of the download/decompress time for each new
release, would.
But thanks in any case :)