Closed boxbox2 closed 1 month ago
RuntimeError: Given groups=1, weight of size [64, 3, 3, 3], expected input[1, 256, 256, 256] to have 3 channels, but got 256 channels instead same... @sjtuplayer
Add more detials, when I run this command: CUDA_VISIBLE_DEVICES=0 python main.py --spatial_encoder_embedding --data_enhance --base configs/latent-diffusion/txt2img-1p4B-finetune-encoder+embedding.yaml -t --actual_resume /home/ai/projects/anomalydiffusion/models/ldm/text2img-large/model.ckpt -n test --gpus 0 --init_word anomaly --mvtec_path=/home/ai//mvtec_anomaly_detection
@sjtuplayer Thanks, Any help would be appreciated!
Maybe you should change batch size from 1 to a number larger than 1(e.g., 2 or 4). I think this is caused by some torch.squeeze() operation.