Closed Kosei1227 closed 1 month ago
Hey,
please check this line in modeling_langbridge.py https://github.com/kaistAI/LangBridge/blob/170e00d8ca90eb4f2e033a91461be582b6f34651/langbridge/modeling_langbridge.py#L51
The code explicitly instantiates the MT5EncoderModel class from HF, since there's no auto classes for EncoderModel (unlike how there's AutoModelForCausalLM etc.).
I would try modifying that part to T5EncoderModel.
Seems like the error is happening because the "else" cause is triggered in the code I shared to you, where it just instantiates the AutoModel class which loads the whole model, including the decoder.
Thank you so much! After additional debugging, I could gain excellent results with our own encoder model. I appreciate your helps.
Hi, Thank you for your assistance so far. I greatly appreciate your help with this research project.
Our research team is thinking about using different encoder-decoder models, called
castorini/afriteva_v2_large
and run the following training script.But, we got the following error. ''' raise ValueError(f"You have to specify either {err_msg_prefix}input_ids or {err_msg_prefix}inputs_embeds") ValueError: You have to specify either decoder_input_ids or decoder_inputs_embeds ''' The exact same error is observed with
google/t5-large-lm-adapt
and other models. I think that in the process of settingsDKYoon/mt5-large-lm-adapt
, you somehow might disable(?) decoders.Could you share some tips/instructions to use different encoder-decoder models?
Thank you so much