Open AnkitSinha123 opened 8 months ago
I have the same error on finishing of Epoch 5:
Average Peak memory 12120.32MiB Epoch 5: 100%|██████████████████████| 101/101 [01:03<00:00, 1.59it/s, loss=0.0902, v_num=0, train/loss_simple_step=0.0691, train/loss_vlb_step=0.000233, train/loss_step=0.0691, global_step=599.0, train/loss_simple_epoch=0.144, train/loss_vlb_epoch=0.00187, train/loss_epoch=0.144]Epoch 5, global step 599: val/loss_simple_ema was not in top 1 Epoch 5: 100%|██████████████████████| 101/101 [01:04<00:00, 1.58it/s, loss=0.0902, v_num=0, train/loss_simple_step=0.0691, train/loss_vlb_step=0.000233, train/loss_step=0.0691, global_step=599.0, train/loss_simple_epoch=0.144, train/loss_vlb_epoch=0.00187, train/loss_epoch=0.144] Saving latest checkpoint...
Traceback (most recent call last):
File "main.py", line 781, in {loader_name}()
method defined to run Trainer.{trainer_method}
.")
pytorch_lightning.utilities.exceptions.MisconfigurationException: No test_dataloader()
method defined to run Trainer.test
.
Somebody has a solution here: https://github.com/Lightning-AI/pytorch-lightning/discussions/11437
you need to pass in the datamodule to trainer.test.
I do not know how to do it.
@surfingnirvana
I found the solution,
In the main file just change
trainer_kwargs["max_steps"] = trainer_opt.max_steps --> trainer_kwargs["max_steps"] = opt.max_steps
It will work
Thank you it works!
Epoch 1: 50%|▍| 201/404 [01:29<01:30, 2.25it/s, loss=0.116, v_num=0, train/loss_simple_step=0.419, train/loss_vlb_step=0.00246, tra Saving latest checkpoint...
Can you please help me with the issues