Closed Chriszhangmw closed 8 months ago
after I kill the process related to sqlite, run again, but get another error:
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks... To disable this warning, you can either:
tokenizers
before the fork if possibleI would delete everything from paperetl/models
and try again. The database locked error would only happen if another process was accessing it. I would make sure there isn't multiple instances of the paperetl process running.
the size is 0
Do you have data in paperetl/data? It seems like nothing was processed.
Closing this issue due to inactivity. Please re-open or open a new issue if there are further questions.
!python -m paperetl.file paperetl/file/data paperetl/models
get the following error:
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks... To disable this warning, you can either:
tokenizers
before the fork if possible