Closed MattJeanes closed 2 years ago
i have the same issue but with gpt2
Thank you @MattJeanes for a detailed analysis and thank you @AbdelrhmanNile for confirming the issue. I created a PR that I think solves the issue.
That's great to hear! Thank you I'll be sure to give it a go when merged 😁
Hey, not sure if this is a bug or my input training file (
train.txt
) is not formatted correctly I cannot get it to output the new line character in the results. The format of the data set is a conversation comprised of around 230,000 messages and I have tried the following formats:Single new line between messages:
Explicit new line character after end of message (resulted in escaped new line in output data):
Double new line between messages:
Using end of text marker (NOTE: split every 10,000 messages): This is the format I have used for GPT-2 training before which has worked
All of these result in the output text being all on one line after some time training, for example:
Person 1: some messagePerson2: another messagePerson3: one more message
. Another model I have works correctly, so I don't believe it is the way I am inferencing on the completed model.Here is the code I use to train the model:
And here is the code I am using to test inference:
So is there something I am doing wrong or is this an issue with Happy Transformers? Thanks