Closed xbox002000 closed 9 months ago
I have the same issue. Have you found a solution yet?
NO~
Same error
Same Error ! Where is the smart genius to help us ? :)
CUDA SETUP: Loading binary C:\Users\Utilisateur\Documents\Kohya\kohya_ss\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cuda116.dll...
use 8-bit Adam optimizer
running training / 学習開始
num train images * repeats / 学習画像の数×繰り返し回数: 1700
num reg images / 正則化画像の数: 0
num batches per epoch / 1epochのバッチ数: 850
num epochs / epoch数: 1
batch size per device / バッチサイズ: 2
total train batch size (with parallel & distributed & accumulation) / 総バッチサイズ(並列学習、勾配合計含む): 2
gradient accumulation steps / 勾配を合計するステップ数 = 1
total optimization steps / 学習ステップ数: 850
Traceback (most recent call last):
File "C:\Users\Utilisateur\Documents\Kohya\kohya_ss\train_network.py", line 573, in
I get the same error too
================================================================================
CUDA SETUP: Loading binary D:\AI 繪圖\lora-scripts-main\lora-scripts-main\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cuda116.dll...
use 8-bit Adam optimizer
override steps. steps for 20 epochs is / 指定エポックまでのステップ数: 2280
Traceback (most recent call last):
File "D:\AI 繪圖\lora-scripts-main\lora-scripts-main\sd-scripts\train_network.py", line 548, in
So, I gaess I have the same issue as most of the users here:
Traceback (most recent call last):
File "C:\Users\matth\Documents\kohya\kohya_ss\train_network.py", line 573, in
Can someone help us here please? Thank you
CUDA out of memory. Tried to allocate 20.00 MiB (G
I suggest you go into "training paramezters" then "advanced configurations" and check the box "Memory efficient attention" is this helping?
OK I will try it. Thank you
I have solved the problem thank you
I have solved the problem thank you
how ?
I have solved the problem thank you
how ?
Find $train_unet_only = 0 in the document named "train.ps1" and let 0 switch to 1
It worked when I stoped using json data as a Configuration file. But I don't know why....
I have solved the problem thank you
how ?
Find $train_unet_only = 0 in the document named "train.ps1" and let 0 switch to 1
Am I blind? Is the "train.ps1" file in the main folder \kohya\kohya_ss ??
Pls where do I find the Document, I still have errors :(
I have solved the problem thank you
how ?
Find $train_unet_only = 0 in the document named "train.ps1" and let 0 switch to 1
Am I blind? Is the "train.ps1" file in the main folder \kohya\kohya_ss ??
Pls where do I find the Document, I still have errors :(
I am using someone else's modpack, so it's possible that the original files don't include train.ps1. Sorry
I got rid of "Use 8bit adam" and it worked !! It's in the advanced tab :))
就是在我们最后配置的那个里面有一项这个
how to find "8bit Adam"? then delete it? pls help me ^-^
I can't find how to disable 8bit Adam, either. I have scanned everything in the advanced config options, and still getting "returned non-zero exit status 1" when trying to train.
Same trouble here. tried lot of options but the train does not run. Any new idea someone ??
Hey guys, I am also experiencing the same problem on MacOS. Has anyone got a solution yet?
you guy can find Adam 8bit in LoRa>Training>Parameters> Basic > Optimizer > AdamW8bit. I changed option AdamW8bit to AdamW and than I could make my LoRa
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
running training / 学習開始 num train images * repeats / 学習画像の数×繰り返し回数: 1500 num reg images / 正則化画像の数: 0 num batches per epoch / 1epochのバッチ数: 1500 num epochs / epoch数: 1 batch size per device / バッチサイズ: 1 total train batch size (with parallel & distributed & accumulation) / 総バッチサイズ(並列学習、勾配合計含む): 1 gradient accumulation steps / 勾配を合計するステップ数 = 1 total optimization steps / 学習ステップ数: 1500 steps: 0%| | 0/1500 [00:00<?, ?it/s]epoch 1/1 E:\SD\Kohya\kohya_ss\venv\lib\site-packages\torch\utils\checkpoint.py:25: UserWarning: None of the inputs have requires_grad=True. Gradients will be None warnings.warn("None of the inputs have requires_grad=True. Gradients will be None") Error no kernel image is available for execution on the device at line 167 in file D:\ai\tool\bitsandbytes\csrc\ops.cu Traceback (most recent call last): File "C:\Users\xbox0\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\xbox0\AppData\Local\Programs\Python\Python310\lib\runpy.py", line 86, in _run_code exec(code, run_globals) File "E:\SD\Kohya\kohya_ss\venv\Scripts\accelerate.exe__main__.py", line 7, in
File "E:\SD\Kohya\kohya_ss\venv\lib\site-packages\accelerate\commands\accelerate_cli.py", line 45, in main
args.func(args)
File "E:\SD\Kohya\kohya_ss\venv\lib\site-packages\accelerate\commands\launch.py", line 1104, in launch_command
simple_launcher(args)
File "E:\SD\Kohya\kohya_ss\venv\lib\site-packages\accelerate\commands\launch.py", line 567, in simple_launcher
raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd)
subprocess.CalledProcessError: Command '['E:\SD\Kohya\kohya_ss\venv\Scripts\python.exe', 'train_network.py', '--enable_bucket', '--pretrained_model_name_or_path=E:/SD/stable-diffusion-webui/models/Stable-diffusion/realisticVisionV13_v13.safetensors', '--train_data_dir=E:/SD/lora data/image', '--resolution=512,512', '--output_dir=E:/SD/lora data/model', '--logging_dir=E:/SD/lora data/log', '--network_alpha=1', '--save_model_as=safetensors', '--network_module=networks.lora', '--text_encoder_lr=5e-5', '--unet_lr=0.0001', '--network_dim=8', '--output_name=last', '--lr_scheduler_num_cycles=1', '--learning_rate=0.0001', '--lr_scheduler=cosine', '--lr_warmup_steps=150', '--train_batch_size=1', '--max_train_steps=1500', '--save_every_n_epochs=1', '--mixed_precision=fp16', '--save_precision=fp16', '--seed=1234', '--cache_latents', '--bucket_reso_steps=64', '--mem_eff_attn', '--gradient_checkpointing', '--xformers', '--use_8bit_adam', '--bucket_no_upscale']' returned non-zero exit status 1.