Closed YsylviaUC closed 1 year ago
Hi @YsylviaUC , Thank you for your interest in our work and for reporting this!
I just fixed that, and set input_ids
to be the labels when labels are not specified.
But I am wondering how did it happen, because it doesn't happen in my scripts (that is, labels
are always provided).
Which model are you using?
Are you using --predict_with_generate
by accident?
Best, Uri
Yes! I used --predict_with_generate by accident and I find this mistake just now. Thank you so much!!
Great!
It's important not to use --predict_with_generate
when saving the datastore, because we want the datastore to be saved according to the training labels, not to the random/generated labels.
Let me know if you have any more questions, Uri
Great! It's important not to use
--predict_with_generate
when saving the datastore, because we want the datastore to be saved according to the training labels, not to the random/generated labels.Let me know if you have any more questions, Uri
Hi~,I have two questions:
3.How could I get the example not just the representation or vectors? Because I am curious about the retrieved example and wanna know more details ;D
Hi @YsylviaUC , Good catch!
This is a flag that we used to run some experiments on randomly deciding on whether or not to perform a kNN search, similarly to some of the experiments with kNN-LM in the paper, which allowed to measure some points between kNN-LM with 0% fraction of saved searches, and the base LM that has 100% saved searches:
So for example, --knn_drop=0.3
meant that 30% of the timesteps used the standard LM, and 70% of the time steps used kNN.
Eventually, we deleted the code for this functionality because we believed that it will not be useful for most users, but accidentally kept the flags, which we should indeed delete.
Thank you for noticing this! Uri
Hi~, I copied kmm_lm.py files into my project and used _knn_wrapper.break_into( my model)_and tried to save a datastore as what instructions do in KNN-MT, but unfortunately, I met some problems when running the code: