Limiting the number of Triples being cached when importing a Sparql EndPoint

8 views
Skip to first unread message

Maatary Okouya

unread,
Apr 4, 2017, 5:33:59 PM4/4/17
to TopBraid Suite Users
Hi, 

I had to substantially increase the Java Heap of Topbraid, to successfully add a sparql end point to a database running on my machine. 


TBC was crashing depending of the amount of data that i loaded on my triple store. It seems that TBC by default, load the all content coming from the TripleStore endpoint. 

Is there a way to limit that, or else to stop TBC from Caching ?


Also maybe the best would be to describe, what is the point of importing a sparql endpoint, as opposed to using SERVICE. What is the original intent of that functionality ? 

Basically could it be that it is to import a model in memory right ? and work with it as if you had an RDF file ?

Holger Knublauch

unread,
Apr 4, 2017, 8:23:25 PM4/4/17
to topbrai...@googlegroups.com
Hi Maarary,

I assume you are referring to the SPARQL Endpoint graphs stored via .sparqlep files and created via "Create Connection File for SPARQL Endpoint"?

That feature is allowing users to treat any (remote) SPARQL endpoint as a normal RDF graph. Basically each find SPO query is turned on the fly into a corresponding SPARQL SELECT ?s ?p ?o query. It is not optimized at all and not recommended for serious work due to its architecture. Among others it also cannot handle blank nodes correctly because the SPARQL protocol doesn't support blank node identity.

If you ask something like ?s ?p ?o (with all three nodes as wildcards) it will indeed end up loading the whole SPARQL end point into a temporary data structure. For medium-sized to large databases this is almost certainly not what you want.

To offer advice on alternatives, you may want to tell us more about what you are trying to do. If you just want to query a remote end point for certain data triples, you could switch to using SERVICE. You could embed SERVICE into SPIN rules or magic properties. If you use SERVICE from SPIN rules, you could perform the "inferences" to download the required triples as a batch job.

Regards,
Holger
--
You received this message because you are subscribed to the Google Group "TopBraid Suite Users", the topics of which include the TopBraid Suite family of products and its base technologies such as SPARQLMotion, SPARQL Web Pages and SPIN.
To post to this group, send email to topbrai...@googlegroups.com
---
You received this message because you are subscribed to the Google Groups "TopBraid Suite Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to topbraid-user...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply all
Reply to author
Forward
0 new messages