Open YanLiang1102 opened 7 years ago
using a tf-idf to find the most important actors in the db that have not be coded yet.
And the ner result with our arabic nouns is imported into a tabled called spacyNer on hanover mongodb
And the output of the data is store on portland. This will be the data we use to build the fast coding interface, key is the arabic entity ,value is the first 5 snetence id that contain this word And the data is stored in mongodb on portland, table is called fast_per_entities and fast_org_entities
Since the wikipedia library is in python, need to use flask to make a python api that the web can make a call to, and for flask python api this is a good tutorial. https://blog.miguelgrinberg.com/post/designing-a-restful-api-with-python-and-flask