Open c1a1o1 opened 3 years ago
Hi! Have you tried to train this tag by yourself? It wouldn't cost much for either time and GPU for a single-tag HiSD model. (maybe several hours on a single 1080Ti.) If you want to manipulate a single image only, you can also try recent CLIP-guided method like StyleCLIP. Please let me know if there are any problems when you reproduce HiSD.
F:\ProgramData\Anaconda3\python.exe F:/work/mingxingshipin/HiSD-main/core/train.py
F:\work\mingxingshipin\HiSD-main\core\utils.py:47: YAMLLoadWarning: calling yaml.load() without Loader=... is deprecated, as the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full details.
return yaml.load(stream)
Traceback (most recent call last):
File "F:/work/mingxingshipin/HiSD-main/core/train.py", line 66, in
Process finished with exit code 1
Could you show me which config file you used in you reproduction?
image_save_iter: 1000 # How often do you want to save output images during training snapshot_save_iter: 10000 # How often do you want to save trained models log_iter: 10 # How often do you want to log the training stats
total_iterations: 600000
batch_size: 8
num_workers: 1
weight_decay: 0
beta1: 0
beta2: 0.99
init: kaiming
lr_dis: 0.0001
lr_gen_mappers: 0.000001
lr_gen_others: 0.0001
adv_w: 1
sty_w: 1
rec_w: 1
style_dim: 256 noise_dim: 32
discriminators:
channels: [32, 64, 128, 256, 512, 512, 512] extractors:
channels: [32, 64, 128, 256, 512, 512, 512] encoder:
channels: [64, 128, 256] translators:
channels: [64, 64, 64, 64, 64, 64, 64, 64] decoder:
channels: [256, 128, 64] mappers:
pre_channels: [256, 256, 256] post_channels: [256, 256, 256]
name: Mouth_Slightly_Open
tag_irrelevant_conditions_dim: 2
attributes:
-
name: 'with'
filename: ../datasets/Mouth_Slightly_Open_with.txt
-
name: 'without'
filename: ../datasets/Mouth_Slightly_Open_without.txt
input_dim: 3 # number of image channels new_size: 256 # first resize the shortest image side to this size crop_image_height: 256 # random crop image of this height crop_image_width: 256 # random crop image of this width
The config file you used need 2x1080Ti. You can change 'new_size', 'crop_image_height' and 'crop_image_weight' in "data options" to 128 or use the "celeba-hq.yaml" to reduce the need of GPU memory.
arguments are located on different GPUs
3090 is not ok???
3090 has 24GB GPU memory, which is enough for the config file you used. But as you comment: "" RuntimeError: CUDA out of memory. Tried to allocate 64.00 MiB (GPU 0; 11.00 GiB total capacity; 8.14 GiB already allocated; 7.74 MiB free; 8.26 GiB reserved in total by PyTorch) "" It seems that you use a 11GB GPU. Maybe you can check if you use the correct GPU.
Can you elease a checkpoint of Open the mouth tag?