Open JervenBolleman opened 3 years ago
See #56 -- "POST quads" (which is roughly "normal HTTP operations on a dataset").
IMHO such functionality is not so easy to standardize because there are many options
vendor:sortOrder vendor:PGSO
above), entity id bit-size, reasoning mode, enabled plugins including secondary indexing, etc etc. Eg see GraphDB repo configconfiguration.ttl
should extend existing RDF-based configuration mechanisms:
Like most databases, GraphDB includes a variety of tools for loading data that are appropriate for different situations:
In addition, GraphDB can push changes to Kafka, and a future version will have ingest from Kafka.
Why?
There is a lot of public data available as RDF but not always available via SPARQL endpoint. It would be nice if the scripts to bulk load such RDF into a store can be shared between vendors. This way a data provider can easily tell people how to load larger datasets.
Previous work
Proposed solution
A single executable (per vendor/sparql db) that consumes two files. The first configures the database connection or the whole database even. The second contains a description of the files (IRIs) to load.
A description of the files to load should be in RDF for extensibility .
The loader program instructs the sparql store how to actually load the data. This is vendor dependant logic and may take extra information regarding the infrastructure into account.
The configuration file will probably be completely custom to the vendor/product.
Ideas giving information regarding the nature of the IRIs to be loaded and how the database/loader should interact with them.
Considerations for backward compatibility
None