Janusgraph as a productive environment for hundreds of millions of nodes #3781
Replies: 2 comments
-
JanusGraph can handle more than quintillion edges and half the amount of vertices. Thus, hundreds of millions isn't that big graph. |
Beta Was this translation helpful? Give feedback.
-
Also important to note is that JanuGraph OLAP operations depend on Apache Spark. For OLAP operations to be efficient, the entire graph will have to be temporarily present in memory as java objects, which might be TB's of cluster RAM in your case (assuming you need shuffling because the graph is globally connected). Also, a cql storage backend will perform much better than HBase for OLAP operarions because of the data partition sizes they use. |
Beta Was this translation helpful? Give feedback.
-
I work in a company which holds hundreds of millions of records for OLAP and other uses currently implemented on AWS Neptune.
It's fair to say that the costs of this operation is HUGE. So i'm currently evaluating different in-house alternatives to aliviate costs and posible implement optimizations for our data structure, could Janusgraph handle efficiently this size of dataset and operate the amount of nodes we currently use, what known limits Janusgraph has?
Beta Was this translation helpful? Give feedback.
All reactions