Hi All,
We are currently facing an issue where there is too many edges (~1m) from node hence it either takes a long time or it doesn’t get back at all and UI cannot handle (does not need to handle the heavy load) . We are looking at filtering based on a edge property that is the count of edges to another node ( there may be many edges to another node ) and sort the outgoing nodes by the max count of edges from the primary node and filter it (50-100) nodes. We are facing some performance issues with this as the filter is on the out going node and not the edges itself. Any easier way to handle this.
--
You received this message because you are subscribed to the Google Groups "JanusGraph users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to janusgraph-use...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/janusgraph-users/ed6ab203-7edc-4ce9-acd6-d332b9cd33dd%40googlegroups.com.
\,,,/
(o o)
-----oOOo-(3)-oOOo-----
plugin activated: janusgraph.imports
plugin activated: tinkerpop.server
plugin activated: tinkerpop.utilities
plugin activated: tinkerpop.hadoop
plugin activated: tinkerpop.spark
plugin activated: tinkerpop.tinkergraph
gremlin> graph = JanusGraphFactory.open("inmemory")
==>standardjanusgraph[inmemory:[127.0.0.1]]
gremlin> GraphOfTheGodsFactory.loadWithoutMixedIndex(graph,true)
==>null
gremlin> g = graph.traversal()
==>graphtraversalsource[standardjanusgraph[inmemory:[127.0.0.1]], standard]
gremlin> a = g.V().bothE().id().next()
18:03:37 WARN org.janusgraph.graphdb.transaction.StandardJanusGraphTx - Query requires iterating over all vertices [()]. For better performance, use indexes
==>35u-368-6c5-38w
gremlin> a.getClass()
==>class org.janusgraph.graphdb.relations.RelationIdentifier
gremlin> a.getInVertexId()
==>4208
gremlin> a.getOutVertexId()
==>4112
gremlin> g.V().inE().group().by(id().map{it->it.get().getOutVertexId()}).by(count()).unfold()
18:09:38 WARN org.janusgraph.graphdb.transaction.StandardJanusGraphTx - Query requires iterating over all vertices [()]. For better performance, use indexes
==>4240=3
==>8208=1
==>4112=4
==>4120=5
==>4232=4
gremlin>
Hi All,
We are currently facing an issue where there is too many edges (~1m) from node hence it either takes a long time or it doesn’t get back at all and UI cannot handle (does not need to handle the heavy load) . We are looking at filtering based on a edge property that is the count of edges to another node ( there may be many edges to another node ) and sort the outgoing nodes by the max count of edges from the primary node and filter it (50-100) nodes. We are facing some performance issues with this as the filter is on the out going node and not the edges itself. Any easier way to handle this.
g.V().has('groupId', 'grp14').as('p').
inE('post').has('date', between(fromDate, endDate)).outV().as('ip').
dedup('p', 'ip').
store('paths').by(
union(
select('p').valueMap('groupName', 'groupType').project('from'),
valueMap('userId', 'userName').project('to'),
select('p'). inE('post').has('date', between(fromDate, endDate)).as('e').outV().where(eq('ip')).select('e').
union(
values('date').dedup().fold().project('dates'),
values('text').count().project('count')
).fold().project('postLink')
).fold()
)I know I am somewhat biased since I work here, but have you tried testing on ScyllaDB as a backing store? It might be able to scale JanusGraph better.What's your backing store right now?Also, can you describe the hardware you're running on? CPUs/cores per node, RAM and Storage per node? (i.e., to rule out things like running from HDDs [shudder] vs. SSDs, too little CPU horsepower, RAM limits, etc.)
On Thu, Jan 16, 2020, 11:15 PM <vincent2...@gmail.com> wrote:
Hi All,
We are currently facing an issue where there is too many edges (~1m) from node hence it either takes a long time or it doesn’t get back at all and UI cannot handle (does not need to handle the heavy load) . We are looking at filtering based on a edge property that is the count of edges to another node ( there may be many edges to another node ) and sort the outgoing nodes by the max count of edges from the primary node and filter it (50-100) nodes. We are facing some performance issues with this as the filter is on the out going node and not the edges itself. Any easier way to handle this.
--
You received this message because you are subscribed to the Google Groups "JanusGraph users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to janusgra...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/janusgraph-users/ed6ab203-7edc-4ce9-acd6-d332b9cd33dd%40googlegroups.com.
I know I am somewhat biased since I work here, but have you tried testing on ScyllaDB as a backing store? It might be able to scale JanusGraph better.What's your backing store right now?Also, can you describe the hardware you're running on? CPUs/cores per node, RAM and Storage per node? (i.e., to rule out things like running from HDDs [shudder] vs. SSDs, too little CPU horsepower, RAM limits, etc.)
On Thu, Jan 16, 2020, 11:15 PM <vincent2...@gmail.com> wrote:
Hi All,
We are currently facing an issue where there is too many edges (~1m) from node hence it either takes a long time or it doesn’t get back at all and UI cannot handle (does not need to handle the heavy load) . We are looking at filtering based on a edge property that is the count of edges to another node ( there may be many edges to another node ) and sort the outgoing nodes by the max count of edges from the primary node and filter it (50-100) nodes. We are facing some performance issues with this as the filter is on the out going node and not the edges itself. Any easier way to handle this.
--
You received this message because you are subscribed to the Google Groups "JanusGraph users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to janusgra...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/janusgraph-users/ed6ab203-7edc-4ce9-acd6-d332b9cd33dd%40googlegroups.com.
I know I am somewhat biased since I work here, but have you tried testing on ScyllaDB as a backing store? It might be able to scale JanusGraph better.What's your backing store right now?Also, can you describe the hardware you're running on? CPUs/cores per node, RAM and Storage per node? (i.e., to rule out things like running from HDDs [shudder] vs. SSDs, too little CPU horsepower, RAM limits, etc.)
On Thu, Jan 16, 2020, 11:15 PM <vincent2...@gmail.com> wrote:
Hi All,
We are currently facing an issue where there is too many edges (~1m) from node hence it either takes a long time or it doesn’t get back at all and UI cannot handle (does not need to handle the heavy load) . We are looking at filtering based on a edge property that is the count of edges to another node ( there may be many edges to another node ) and sort the outgoing nodes by the max count of edges from the primary node and filter it (50-100) nodes. We are facing some performance issues with this as the filter is on the out going node and not the edges itself. Any easier way to handle this.
--
You received this message because you are subscribed to the Google Groups "JanusGraph users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to janusgra...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/janusgraph-users/ed6ab203-7edc-4ce9-acd6-d332b9cd33dd%40googlegroups.com.
I know I am somewhat biased since I work here, but have you tried testing on ScyllaDB as a backing store? It might be able to scale JanusGraph better.What's your backing store right now?Also, can you describe the hardware you're running on? CPUs/cores per node, RAM and Storage per node? (i.e., to rule out things like running from HDDs [shudder] vs. SSDs, too little CPU horsepower, RAM limits, etc.)
On Thu, Jan 16, 2020, 11:15 PM <vincent2...@gmail.com> wrote:
Hi All,
We are currently facing an issue where there is too many edges (~1m) from node hence it either takes a long time or it doesn’t get back at all and UI cannot handle (does not need to handle the heavy load) . We are looking at filtering based on a edge property that is the count of edges to another node ( there may be many edges to another node ) and sort the outgoing nodes by the max count of edges from the primary node and filter it (50-100) nodes. We are facing some performance issues with this as the filter is on the out going node and not the edges itself. Any easier way to handle this.
--
You received this message because you are subscribed to the Google Groups "JanusGraph users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to janusgra...@googlegroups.com.