Closed Maxwell-Lyu closed 2 years ago
Thanks @Maxwell-Lyu ! Fixed! Let me know if you have any problems or questions.
Thanks! Closing this issue now. Glad I can help~
Sorry to disturb you.
I run kNN-MT with default hyper-prams, but it got the error: OverflowError: out of range integral type conversion attempted
. It seems the default max_length hyper-pram for decoding process is small or another reason.
How can I fix this error? Thank you!
Hi @vhientran , Thank you for your interest in our work!
This is a closed issue, can you please open a new one, and provide the details of what you ran exactly, what is the full error and stack trace?
Thanks, Uri
Hi @Maxwell-Lyu and @urialon , Sorry for disturbing you. I only wonder how we can set or select the value of the hyper-parameter dstore_size effectively and suitably?
Hi @vhientran ,
dstore_size
is the total number of tokens in your training set. You can find this number by starting an evaluation of any model on the train split,
Using the command line here:
https://github.com/neulab/knn-transformers#step-1-evaluating-the-base-language-model
But using --eval_subset train
instead of validation
.
And the number of tokens will be printed by this line:
https://github.com/neulab/knn-transformers/blob/master/run_clm.py#L538
After they are printed, you can stop the run and feed this parameter to a new run.
Best, Uri
Hi @urialon Thank you so much for your detail explanation. It helps me a lot. I will follow your guidance to find dstore_size in my model. Many thanks!
saving-a-datastore-for-knn-mt section in README is missing a proper
dstore_size
parameter.to add
--dstore_size 26565876 \
current version
correct version (maybe)