Open ken2190 opened 1 year ago
set --mixed_precision="fp16" get another error
CalledProcessError: Command '['/usr/bin/python3', '/content/Dreambooth/train.py', '--lora_rank=', '--pretrained_model_name_or_path=', '--pretrained_vae_name_or_path=', '--instance_data_dir=', '--class_data_dir=', '--output_dir=', '--logging_dir=', '--prior_loss_weight=', '--instance_prompt=', '--class_prompt=', '--conditioning_dropout_prob=', '--unconditional_prompt=', '--seed=', '--resolution=', '--train_batch_size=', '--gradient_accumulation_steps=', '--mixed_precision=', '--adam_beta1=', '--adam_beta2=', '--adam_weight_decay=', '--adam_epsilon=', '--learning_rate=', '--learning_rate_text=', '--lr_scheduler=', '--lr_warmup_steps=', '--lr_cosine_num_cycles=', '--ema_inv_gamma=', '--ema_power=', '--ema_min_value=', '--ema_max_value=', '--max_train_steps=', '--num_class_images=', '--sample_batch_size=', '--save_min_steps=', '--save_interval=', '--n_save_sample=', '--save_sample_prompt=', '--save_sample_negative_prompt=']' returned non-zero exit status 2.
I also have the isse , be sure to run all the previous step , including the experimental step
I had a very similar error, which also resolved once I did not skip past the Experimental steps.
i get the below error when i run training cell in colab FineTuning_colab.ipynb also run cell Training parameters and all parameter parsed
No LSB modules are available. Description: Ubuntu 20.04.5 LTS diffusers==0.11.1 lora-diffusion @ file:///content/lora torchvision @ https://download.pytorch.org/whl/cu118/torchvision-0.15.1%2Bcu118-cp39-cp39-linux_x86_64.whl transformers==4.25.1 xformers==0.0.16rc425 2023-04-16 09:29:59.351268: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations. To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags. 2023-04-16 09:30:00.246985: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT
Copy-and-paste the text below in your GitHub issue
Accelerate
version: 0.15.0Accelerate
default config: Not found 2023-04-16 09:30:04.094704: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations. To enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags. 2023-04-16 09:30:04.940115: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT usage: accelerate