facebookresearch / fairseq

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
MIT License
29.78k stars 6.3k forks source link

assert step < max_len, f"{step} < {max_len}" AssertionError: 60 < 60 #5452

Open sevensix617 opened 3 months ago

sevensix617 commented 3 months ago

When I was training a summarization model under fairseq, I encountered the following error in the validation step. How can I solve this problem.

-- Process 1 terminated with the following error: Traceback (most recent call last): File "/home/deep/.conda/envs/torch3.7/lib/python3.7/site-packages/torch/multiprocessing/spawn.py", line 69, in _wrap fn(i, *args) File "/home/deep/.conda/envs/torch3.7/lib/python3.7/site-packages/fairseq/distributed/utils.py", line 328, in distributed_main main(cfg, kwargs) File "/home/deep/PRGEN/train.py", line 181, in main valid_losses, should_stop = train(cfg, trainer, task, epoch_itr) File "/home/deep/.conda/envs/torch3.7/lib/python3.7/contextlib.py", line 74, in inner return func(*args, *kwds) File "/home/deep/PRGEN/train.py", line 314, in train cfg, trainer, task, epoch_itr, valid_subsets, end_of_epoch, File "/home/deep/PRGEN/train.py", line 404, in validate_and_save valid_losses = validate(cfg, trainer, task, epoch_itr, valid_subsets) File "/home/deep/PRGEN/train.py", line 476, in validate trainer.valid_step(sample) File "/home/deep/.conda/envs/torch3.7/lib/python3.7/contextlib.py", line 74, in inner return func(args, kwds) File "/home/deep/PRGEN/trainer.py", line 1037, in valid_step sample, self.model, self.criterion, *extra_kwargs File "/home/deep/PRGEN/src/task/faithful_summary_task.py", line 282, in valid_step metrics = self._inference_with_bleu(self.sequence_generator, sample, model) File "/home/deep/PRGEN/src/task/faithful_summary_task.py", line 332, in _inference_with_bleu gen_out = self.inference_step(generator, [model], sample, prefix_tokens=None) File "/home/deep/.conda/envs/torch3.7/lib/python3.7/site-packages/fairseq/tasks/fairseq_task.py", line 541, in inference_step models, sample, prefix_tokens=prefix_tokens, constraints=constraints File "/home/deep/.conda/envs/torch3.7/lib/python3.7/site-packages/torch/autograd/grad_mode.py", line 27, in decorate_context return func(args, kwargs) File "/home/deep/.conda/envs/torch3.7/lib/python3.7/site-packages/fairseq/sequence_generator.py", line 204, in generate return self._generate(sample, kwargs) File "/home/deep/.conda/envs/torch3.7/lib/python3.7/site-packages/fairseq/sequence_generator.py", line 470, in _generate assert step < max_len, f"{step} < {max_len}" AssertionError: 60 < 60

usefulhyun commented 4 weeks ago

I have the same issue. Did you solve the problem?