Open totaltube opened 1 year ago
what version or codebase of OpenNMT-py did you use? it seems that you have both position_encoding=True and max_relative_position !=0 it is now tested and avoided: https://github.com/OpenNMT/OpenNMT-py/blame/master/onmt/utils/parse.py#L302
what version or codebase of OpenNMT-py did you use? it seems that you have both position_encoding=True and max_relative_position !=0 it is now tested and avoided: https://github.com/OpenNMT/OpenNMT-py/blame/master/onmt/utils/parse.py#L302
Master version on the time of writting. position_encoding: false
With max_relative_position = 20 - converstion goes ok, but with -1 or -2 it fails.
I see that you are trying to convert an encoder-decoder model (_get_model_spec_seq2seq
is in the stack trace), but the converter currently does not handle max_relative_positions: -1
or max_relative_positions: -2
for these models.
I see that you are trying to convert an encoder-decoder model (
_get_model_spec_seq2seq
is the stack trace), but the converter currently does not handlemax_relative_positions: -1
ormax_relative_positions: -2
for these models.
Yes. Ok, just wanted to test these new options for nmt tasks. Anyways, old options works good)
you still can assess your model with regular opennmt-py inference, I am also intrested in such results. we'll add those options in the encoder/decoder config if it makes sense.
At least, I tested it with the following options: add_ffnbias: false, multiquery: true, add_qkvbias: false. I also added other layers to ensure that the model has the same or more parameters. However, it performed worse compared to the standard options.
you still can assess your model with regular opennmt-py inference, I am also intrested in such results. we'll add those options in the encoder/decoder config if it makes sense.
Please add those options to the ctranslate framework so that it can run onmt-py model trained with rotary embeddings. And can you also permit the co-existence of the regular encoder/decoder embeddings (like the sinusoidal one) along with the relative position embedding? Becasue today I found that the onmt's context attention part (encoder-decoder attention) does not support the relative position feature, so enabling the classical position embedding may help in the context attention scenario.
I get this error when
max_relative_positions: -1
ormax_relative_positions: -2