Open ChristLBUPT opened 1 year ago
I am not sure I understand your question correctly. It seems that you are not using {"mask"} and therefore not using verbalizer
on the whole. That way you probably should just use the transformers
library, without wrapping it up with PromptModel
in openprompt.
I want to do prompt tuning for a masked-fill-based T5 model, which has the input format like this:
if I use the template similar to that given by
2.1_conditional_generation.py
, that is:it will automatically assign an
<extra_id_0>
at the corresponding position of{"mask"}
, splitting the original sentence and target sentence with special token</s>,
which results in duplicate<extra_id_0>
s in input sentence, just as follows:I know it is possible to manually add 1 to each extra_id in my dataset, but is it possible to ONLY use the source sentence as input and avoid automatically adding extra ids?