Closed dlorenc closed 3 years ago
Wow! Is the Bug query open for everyone? We should fix this and add it to the readme.
Wow! Is the Bug query open for everyone? We should fix this and add it to the readme.
Yup! It should be publicly readable. It's just an automated import from the GCS bucket.
@dlorenc Need your help with this. I don't have permission. I am trying to update the Big Query with the latest json structure.
@naveensrinivasan - made you bigquery admin, go ahead!
@naveensrinivasan - made you bigquery admin, go ahead!
Thanks.
This probably stopped working due to https://github.com/ossf/scorecard/commit/0eaa4ff3d0dd36cd1d200662e5fc9803cc2fd7be. The BQ import expects newline delimited JSON objects, rather than a well formed JSON file.
This is still failing. Anything I can do to help fix it?
AFAIK the json dump has to be changed.
Dan, see https://github.com/ossf/scorecard/issues/336. I'm working on it right now, but if the failures are blocking work in some way, I can submit a hacky quick-fix solution to unblock this. Let me know.
Unfortunately this is still failing because the Date format in the latest.json is wrong:
Invalid date: '21046-04-21' Field: Date; Value: 21046-04-21
I have a fix in #353
The last two runs have been green.
The cron still isn't importing correctly for example the data for this isn't populated in this tableSELECT * FROM
openssf.scorecardcron.scorecard_latestLIMIT 1000
The old table isn't populated either SELECT distinct(date) FROM
openssf.scorecardcron.scorecardLIMIT 100
the last import in old table is 04-07
So the results are being imported - if you "ORDER BY Date DESC", you'll see the rows. However, turns out that since the output has "CheckResults" instead of "Checks", the checks in BQ are empty.
So the results are being imported - if you "ORDER BY Date DESC", you'll see the rows. However, turns out that since the output has "CheckResults" instead of "Checks", the checks in BQ are empty.
OK, Thanks! So we will need to address the checks
/checkresults
Yes. Working on a PR for that, will send out shortly.
Nice catch!
FYI I manually fixed up the latest.json from the last cron job and recreated the BQ table (and added Date partitionining).
Awesome, thanks Oliver!
The cron stopped importing in biquery:
There's a way to debut this but I forget how right now, you need to use gcloud with the job ids to search for the errors. Opening this bug to track that so I don't forget.