Closed MM-IR closed 6 months ago
To be specific, I am testing on some sentiment transfer tasks, especially in positive transfer tasks with OPT models, I frequently find such issues occurred with several prompts (several positive prompts fail, but I see the same prompt does not fail with other LMs).
Hi! Thanks for reporting this! This seems like an issue caused by exceptionally long sequences. Can you share the code to reproduce this error so we can look into this?
Hi, I work with OPT-IML-1.3b models, and in addition to such phenomena, I find your results are inconsistent with a lot of other implementations, such as initial Huggingface API, and the DeepSpeed wrappers.
I do suggest you to experiment with OPT-IML series.
For the code, I am less likely to think it matters a lot, as I also experiment these prompts (yeah, they are long with100~200 tokens) with other model instances, such as OpenASSISTANT-pythia-12b, LLaMa-7B, etc. They all run smoothly...
Could you check again for OPT-IML instances?
Could you check again for OPT-IML instances?
Sorry to bother again, can you provide a concrete prompt example that can reproduce the error on opt-iml-1.3b
? We tried some simple examples before but did not find errors.
Closing this issue as stale as there has been no discussion in the past 3 months.
If you are still experiencing the issue you describe, feel free to re-open this issue.
Hi, when using the model.generate API, I just run into this issue (200 dataset samples, stuck at 30%). If possible, could you provide any suggestions? This is really weird, as a rare case to this particular NL prompt, where your api runs smoothly with other prompt cases in my current tests.
Thanks in advance!