usc-isi-i2 / dig-etl-engine

Download DIG to run on your laptop or server.
http://usc-isi-i2.github.io/dig/
MIT License
101 stars 39 forks source link

Can you outline the process how to extract information from structured data sources like csv file which had been modeled using KARMA? #279

Open llFireHawkll opened 5 years ago

llFireHawkll commented 5 years ago

I have many datasets in the form of csv files I wanted to know what is the process to create a KG using the semantic modeling from KARMA and then integrating it with the DIG Tool? Is it even possible to extract information from csv file as in the documentation part the only thing I found to be mentioned is about extracting information using the webpages? I want you to please provide me with the correct steps I need to follow to model data using KARMA and then importing it to DIG so that I can extract information from csv files or tables. Thanks

briantdu777 commented 5 years ago

@llFireHawkll Just curious, were you ever able to figure out this process?

I was able to upload the JSON file that was exported from KARMA by adding extra fields like "url", "doc_id", and "raw_content", then convert it into a JSON Lines file, so that when the file is uploaded to DIG, the em modules can know what to do with the file, but I don't think that is necessarily helpful persay because the actual fields that we care about form the JSON file aren't even being parsed and taken into consideration for the KG.

Thanks in advance!