--
---
You received this message because you are subscribed to the Google Groups "DigitalNZ" group.
To unsubscribe from this group and stop receiving emails from it, send an email to digitalnz+...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
Conal,
I recommend that you focus on the proxy model for now. From memory, the terms of use don't allow caching of responses.
Once the proof of concept is built and functional, you can then work with the DNZ team to go back to content providers and ask for a more liberal licence.
Being part of the LOD cloud would be pretty neat! Good luck.
In general you are not discouraged from caching at all, but you are allowed to cache data for up to 30 days if you have a good technical reason.
Hi Conal,
Thanks for proposing this as it sounds really interesting. A couple of thoughts from my side:
· I can confirm that the terms and conditions do allow for caching up to 30 days to support localised applications
· The intent behind the caching is to support discreet localised experiences, as opposed to supporting the complete download of the DigitalNZ held data. That’s not to say that a full download of the corpus could not be valuable, simply that the systems and agreements with partners are not currently designed with that in mind
· The reason for caching is that this is a real time system with data changing constantly, and our partners need assurance that the changes they make will flow through (that modifications, additions, and even complete deletion of records will be respected)
· If you are wanting to pursue this then I’d recommend you start with a smaller prototype of discreet data, and we’d certainly be keen to see the results
As an aside, we’ve looked at the SPARQL options previously and came to the conclusion that the metadata quality may not be good enough yet to make this a really valuable service en mass. Certainly our existing infrastructure couldn’t cope with heavy SPARQL queries which is why we haven’t gone down that route. The concepts API is our first foray into the linking of data and there will be more to come on this. It’s certainly something we continue to work on.
You’re right that the concepts themselves don’t point to related records, but it works the other way, you can query the records API for a matching concept. More documentation is here:
http://digitalnz.github.io/supplejack/api_usage/concepts-api.html
If there are specific question on this that we can help with just give me a bell.
Andy
Hi Conal,
Thanks for proposing this as it sounds really interesting. A couple of thoughts from my side:
· I can confirm that the terms and conditions do allow for caching up to 30 days to support localised applications
· The intent behind the caching is to support discreet localised experiences, as opposed to supporting the complete download of the DigitalNZ held data. That’s not to say that a full download of the corpus could not be valuable, simply that the systems and agreements with partners are not currently designed with that in mind
As an aside, we’ve looked at the SPARQL options previously and came to the conclusion that the metadata quality may not be good enough yet to make this a really valuable service en mass.
Certainly our existing infrastructure couldn’t cope with heavy SPARQL queries which is why we haven’t gone down that route.
The concepts API is our first foray into the linking of data and there will be more to come on this. It’s certainly something we continue to work on.
You’re right that the concepts themselves don’t point to related records, but it works the other way, you can query the records API for a matching concept.
Hi Conal,
A couple of quick thoughts:
· Yes, I definitely agree that publishing bulk datasets would be better, and that the current APIs were not designed with that in mind. There is a technical component to this, but keep in mind also that DNZ is an aggregation service so we need to respect the copyright status of the data and the rights that owners provide… even while we advocate for more permissible uses that would allow bulk download
· If a record has any concepts attached you can get to that data in the concept_ids field e.g. http://api.digitalnz.org/records/35111213.json?api_key=your-key
· I think I’ll refrain from getting further into the merits and practicality of SPARQL:-) but perhaps we should have that discussion over a drink at the National Digital Forum!
· If a record has any concepts attached you can get to that data in the concept_ids field e.g. http://api.digitalnz.org/records/35111213.json?api_key=your-key
Hi again!
No, we don’t have anything analogous to Sets… but it’s a nice idea!
And we haven’t yet taken the step to identify the type of relationships. The basic approach from the beta has been to simply accept the relationships that have been determined by our contributing content partners i.e. if a National Library item record is connected to a National Library authority file, then that’s what we record. So if you are not seeing an assertion then it means it has not yet been connected by the content partner. We’re just scratching the surface in terms of possible relationships, and it is a very simple start.
A.
From: digi...@googlegroups.com [mailto:digi...@googlegroups.com]
On Behalf Of Conal Tuohy
Sent: Thursday, 24 March 2016 3:07 p.m.
To: digi...@googlegroups.com
Subject: Re: [DigitalNZ] Converting Digital NZ data to Linked Open Data
On 24 March 2016 at 10:32, Andy Neale <Andy....@dia.govt.nz> wrote:
--
Hi again!
No, we don’t have anything analogous to Sets… but it’s a nice idea!
And we haven’t yet taken the step to identify the type of relationships. The basic approach from the beta has been to simply accept the relationships that have been determined by our contributing content partners i.e. if a National Library item record is connected to a National Library authority file, then that’s what we record.