bcgov / entity

ServiceBC Registry Team working on Legal Entities
Apache License 2.0
23 stars 59 forks source link

Run UAT Rejection Tests #5094

Closed LJTrent closed 4 years ago

LJTrent commented 4 years ago

Description:

The rejection uat is a python script that runs the auto-analyze against rejected NRs in postgres. It requires use of namex-api and synonyms-api.

https://github.com/bcgov/namex/tree/master/jobs/rejection-uat

This is currently run against the production Namex Postgres database and only includes BC Corps so the query can be tweaked to whatever other entities you want to include.

The original examiner results and the auto-analyze results are stored in a table called like this: uat = UatResults() uat.id = data_dict['id'] uat.nr_num = data_dict['nr_num'] uat.nr_state = data_dict['nr_state'] uat.choice = data_dict['choice'] uat.name = data_dict['name'] uat.name_state = data_dict['name_state'] uat.decision_text = data_dict['decision_text'] uat.conflict_num1 = data_dict['conflict_num1'] uat.conflict1 = data_dict['conflict1'] uat.result_state = data_dict['result_state'] uat.result_decision_text = data_dict['result_decision_text'] uat.result_conflict_num1 = data_dict['result_conflict_num1'] uat.result_conflict1 = data_dict['result_conflict1'] uat.result_response = data_dict['result_response'] uat.result_duration_secs = data_dict['result_duration_secs']

All columns that start with result are the auto-analyze results all other columns are the original examination.

Once the script is run, the results are exported to a spreadsheet and given to Genevieve to review. She reviews any differences in decision. That may entail looking at conflict names to ensure they are still a conflict (not historical), verifying synonyms, word classification, consent words were necessary and providing feedback to Arturo for missing rules.

Of the ones we ran. we did not have any the auto-analyze approved that examiners rejected. We did have ones the we conflicted but examiners approved. The set was 185. It is time-consuming to run the script depending on how common a name is and how many conflicts. we retrieve all conflicts and score them.

We are missing some rules and we also are in the process of tweaking the score.

Details

jdyck-fw commented 4 years ago

This estimate was likely wrong, instead it would be closer to 25