Closed MatthewCYM closed 1 year ago
Hi,
There are multiple factors affecting system performance, such as number of processes, your cpu/gpu computing power. More importantly, your way of picking prompts is also another factor, such as the number of steps. In terms of us, we pick prompts when the reward does not have any significant improvements, this may reduce some computing time as well when averaging over random seeds.
Since it is a clarification, I am closing this now.
Thanks for the quick reply. I run the code with
The training takes around 1 day to complete on a single RTX3090, which is much longer than the training time reported in the paper (4 hr). May I ask if this is normal?
I also try to run the code with gpt2 backbone:
The eval accuracy is only 62.5. Have you experimented with GPT2 on the classification task?