Open pawanrawal opened 7 years ago
Hi,
Thanks for your interest, I hope it will be a useful tool and I would like to hear more about your work.
I would appreciate if you can share the schema with me so that I can compare the original schema of datagen and schema of the tables created by he script.
I am using the schema mentioned at https://github.com/anilpacaci/graph-benchmarking/blob/master/snb-interactive-sql/scripts/db_schema.sql
Read-only query latencies are measured by only executing a specific type of query manually (person lookup for point queries, retrieving immediate friends for 1-hop neighbourhood, finding shortest path between two random person for SSSP etc.)
So was only one query with a specific id run per type to get the latencies?
We at Dgraph are trying to reproduce the benchmarks mentioned here https://event.cwi.nl/grades/2017/12-Apaci.pdf and write a blog post comparing Dgraph against the mentioned options. I have some questions. I am specifically interested in comparison against PostgreSQL, Titan and Neo4j.
Would be great if the
load_csv.sql
script can be updated or you can specify the version of the data generator that was used to generate this data.The paper mentions 4 types of query latencies but there are 13 LDBC queries in the benchmark. How are the queries grouped? Is there a framework for evaluating the read-only query performance for Postgres?
Was the data ingestion done after adding the indexes or without them?
In general, some more documentation and steps to reproduce the benchmarks would be very useful.