Closed Edvardoh closed 1 year ago
Trying to take it. Any connection codes?
Hi @jeremyikwuje - thanks for the response. Are you referring to DB connection codes? There is a .zip file with a sample dump from the DB, so you should have all the info you need there to produce the Python script. Let me know if there is anything I might have missed..
@gonzaloaune I think we need sample outputs for the same Runs of the other tables so they can get all the data like jobs, sats_spent, ocr counts, etc?
Got this question from BrainShark: I was already done extracting the main fields, what was only left was refactoring the code and just a little clarification with a certain ‘run_link‘ for the elections(under the Codigos list) and also how to go about the {‘aws_ocr_count‘, ’stak_ocr_count’, ’donut_vqa_count’} which are in the example.json file
Here's a link to the script on Google colab they're building: https://colab.research.google.com/drive/1wooF1GUeWbsijYAjzl0fu4xhGCTg8ubX?usp=sharing
Closing this as we updated the spec to specifically use PySpark/AWS Glue: https://github.com/stakwork/stak-bounties/issues/2
We need to create a custom JSON Report from Stakwork Runs exported from our database.
We have extracted a sample of the Runs table into a csv called query-results.csv, as well as the schema, and an example JSON output.
All of the necessary materials can be found in this zip folder: glue.zip
Deliverables:
Example JSON Format
We also have Ruby on Rails Code which generates the same report, you can use this to help do the same in Python with the csv data.