Closed ericjanto closed 1 year ago
The main things to notice are:
nlp.pipe(batched_text_list)
will have a slight performance boost (running all nlp-ing on the entire Max Havelaar
corpus took ~1.10min as opposed to ~1.30min with my own custom batching). So worth doing at some point but not now