Open euminds opened 1 year ago
I think the below command line output means the training progress is finished when setting max_steps: 800
in aigc_id.yaml
Epoch 0: 80%|▊| 801/1001 [06:35<01:38, 2.02it/s ...
I think the below command line output means the training progress is finished when setting
max_steps: 800
in aigc_id.yamlEpoch 0: 80%|▊| 801/1001 [06:35<01:38, 2.02it/s ...
Another error When I run the Generation bash ./02_start_test.sh ./weights/sd-v1-4-full-ema.ckpt ./infer_images/example_prompt.txt training2023-09-07T09-11-56_celebbasis Follow
Optionally, in ./02_start_test.sh, you may modify the following var as you need:
step_list=(799) # the step of trained '.pt' files, e.g. (99 199 299 399) eval_id1_list=(0) # the ID index of the 1st person, e.g. (0 1 2 3 4) eval_id2_list=(1) # the ID index of the 2nd person, e.g. (0 1 2 3 4) Testing
bash ./02_start_test.sh "./weights/sd-v1-4-full-ema.ckpt" "./infer_images/example_prompt.txt" "traininYYYY-MM-DDTHH-MM-SS_celebbasis" The generated images are under ./outputs/traininYYYY-MM-DDTHH-MM-SS_celebbasis.
I got another error
Traceback (most recent call last):
File "scripts/stable_txt2img.py", line 385, in
Did you free the GPU memory before running the generation code? Typically, your GPU with 24GB memory is sufficient for our code.
I got this error:
Epoch 0: 80%|▊| 800/1001 [06:35<01:39, 2.03it/s, loss=0.0671, v_num=0, train/loss_simple_step=0.0197, train/loss_vlb_step=7.03e-5, train/loss_step=0Epoch 0, global step 799: val/loss_simple_ema was not in top 1 Average Epoch time: 395.59 seconds Average Peak memory 19447.18MiB Epoch 0: 80%|▊| 801/1001 [06:35<01:38, 2.02it/s, loss=0.0671, v_num=0, train/loss_simple_step=0.0197, train/loss_vlb_step=7.03e-5, train/loss_step=0 Saving latest checkpoint...
Traceback (most recent call last): File "main_id_embed.py", line 817, in
trainer.test(model, data)
File "/home/user/miniconda3/envs/celebbasis/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 911, in test
return self._call_and_handle_interrupt(self._test_impl, model, dataloaders, ckpt_path, verbose, datamodule)
File "/home/user/miniconda3/envs/celebbasis/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 685, in _call_and_handle_interrupt
return trainer_fn(*args, **kwargs)
File "/home/user/miniconda3/envs/celebbasis/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 954, in _test_impl
results = self._run(model, ckpt_path=self.tested_ckpt_path)
File "/home/user/miniconda3/envs/celebbasis/lib/python3.8/site-packages/pytorch_lightning/trainer/trainer.py", line 1128, in _run
verify_loop_configurations(self)
File "/home/user/miniconda3/envs/celebbasis/lib/python3.8/site-packages/pytorch_lightning/trainer/configuration_validator.py", line 42, in verify_loop_configurations
__verify_eval_loop_configuration(trainer, model, "test")
File "/home/user/miniconda3/envs/celebbasis/lib/python3.8/site-packages/pytorch_lightning/trainer/configuration_validator.py", line 186, in __verify_eval_loop_configuration
raise MisconfigurationException(f"No
{loader_name}()
method defined to runTrainer.{trainer_method}
.") pytorch_lightning.utilities.exceptions.MisconfigurationException: Notest_dataloader()
method defined to runTrainer.test
.env: I have configured the environment by following these steps:
I used the command conda env create -f environment.yaml to create the environment based on the specifications provided in the environment.yaml file.
Then, I activated the environment using the command conda activate celebbasis. By changing the environment name to "celebbasis," I ensured that I am working within this specific environment.
Regarding the dependencies mentioned:
The line # - -e git+https://github.com/CompVis/taming-transformers.git@master#egg=taming-transformers refers to the "taming-transformers" library. Due to network issues, I independently installed this dependency using the command pip install -e git+https://github.com/CompVis/taming-transformers.git@master#egg=taming-transformers.
Similarly, the line # - -e git+https://github.com/openai/CLIP.git@main#egg=clip refers to the "CLIP" library. You installed this dependency separately as well, using the command pip install -e git+https://github.com/openai/CLIP.git@main#egg=clip.
Next , About Face Alignment I align the images in Img, like id0(1).jpg,...,id0(10).jpg. I think this is a minor issue. Then, I run bash ./01_start_train.sh ./weights/sd-v1-4-full-ema.ckpt and meet the top error
Thanks