Open FleetAdmiral opened 1 year ago
I got the same error
Try to down grade Fairseq to the previous version.
You can try the following lines:
!pip install torch==1.12.1+cu113 torchvision==0.13.1+cu113 torchaudio==0.12.1 --extra-index-url https://download.pytorch.org/whl/cu113
then
!pip install fairseq==0.12.2
it solved the issue, the problem was with the new version of torch that is installed on Colab, so installing the previous version of torch solves the issue.
@OmarAshrafFathy Thankyou for this. This error gave me a shock on the piece of code which we hadn't touched since months. 😂
Love you @OmarAshrafFathy you saved me
The model I trained with fairseq 0.12.2 and torch 2.1.0 also encountered this situation. If the torch version is reduced, does the model still need to be retrained? @OmarAshrafFathy !thank you
@boolmriver No, you don't need to retrain the model again.
Sorry I don't think it's optimal to fix this problem with a downgraded version. It looks like a problem with the higher version of torch? Is there a solution to this problem
I have a simpler solution , which is to skip _transformer_encoder_layer_fwd by setting can_use_fastpath to False at generate, but that doesn't look good
This issue seems to be fixed on the latest main
branch.
Some background:
I'm working on a translation problem where I am able to get through the fairseq-preprocess and fairseq-train but during the process of fairseq-generate, the operation fails in the middle.
I have not found any mention of this error message online as an issue or in any documentation.
What I've attempted from my end:
Reducing train/test size. Increase the train and/or test size. Making sure the test dataset has no unknown token. I'm a novice so this may look elementary, but I'd really appreciate it if you can help me out here.
This is the error that is then presented: