Closed guozix closed 11 months ago
Thank you for your interest! Please refer to this issue for the explanation. Basically, this is GPT-2's way of representing "space". Let me know if you have any other question. I am closing this issue for now because it's a clarification question.
Thanks for your patient reply! greatly appreciated!
I understand that the involvement of \u0120
is naturally caused by the construction of the BPE tokenizer based on pre-train data.
I have another three questions:
\u0120
-prepended" words a conventional design or just an empirical design for RLPrompt?\u0120
-prepended" verbalizers aims for the result of, take sentiment classification as example, "\<sentence> It was\<space>great" instead of "\<sentence> It wasgreat" ?\u0120
-prepended" verbalizers? Can it achieve comparable results to current design of verbalizers?
Thanks for releasing the code!
I have some questions about the prepended special character \u0120 in verbalizers: