Open zarigata opened 3 months ago
Hey, please dont use DirectML for this GPU. You can run it with ZLUDA. Its much faster and has a better Vram management. I have the same card too and no issues. You find my AMD Automatic1111 with ZLUDA Guide here: https://github.com/CS1o/Stable-Diffusion-Info/wiki/Installation-Guides
Hey, please dont use DirectML for this GPU. You can run it with ZLUDA. Its much faster and has a better Vram management. I have the same card too and no issues. You find my AMD Automatic1111 with ZLUDA Guide here: https://github.com/CS1o/Stable-Diffusion-Info/wiki/Installation-Guides
ive folowed this guide and yet no sucess.... maybe there is a new update in Zluda?
`stderr: error: subprocess-exited-with-error
Preparing metadata (pyproject.toml) did not run successfully. exit code: 1
[17 lines of output]
meson setup C:\Users\zarigata\AppData\Local\Temp\pip-install-rld2l6i6\scikit-image_1fead003d3cc4b6db02665cc03107916 C:\Users\zarigata\AppData\Local\Temp\pip-install-rld2l6i6\scikit-image_1fead003d3cc4b6db02665cc03107916.mesonpy-3ge7pess -Dbuildtype=release -Db_ndebug=if-release -Db_vscrt=md --native-file=C:\Users\zarigata\AppData\Local\Temp\pip-install-rld2l6i6\scikit-image_1fead003d3cc4b6db02665cc03107916.mesonpy-3ge7pess\meson-python-native-file.ini The Meson build system Version: 1.5.1 Source dir: C:\Users\zarigata\AppData\Local\Temp\pip-install-rld2l6i6\scikit-image_1fead003d3cc4b6db02665cc03107916 Build dir: C:\Users\zarigata\AppData\Local\Temp\pip-install-rld2l6i6\scikit-image_1fead003d3cc4b6db02665cc03107916.mesonpy-3ge7pess Build type: native build Project name: scikit-image Project version: 0.21.0
..\meson.build:1:0: ERROR: Unable to detect linker for compiler clang -Wl,--version
stdout:
stderr: clang: warning: unable to find a Visual Studio installation; try running Clang from a developer command prompt [-Wmsvc-not-found]
clang: error: unable to execute command: program not executable
clang: error: linker command failed with exit code 1 (use -v to see invocation)
A full log can be found at C:\Users\zarigata\AppData\Local\Temp\pip-install-rld2l6i6\scikit-image_1fead003d3cc4b6db02665cc03107916.mesonpy-3ge7pess\meson-logs\meson-log.txt [end of output]
note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed
Encountered error while generating package metadata.
See above for output.
note: This is an issue with the package mentioned above, not pip. hint: See above for details.
Press any key to continue . . .`
Hey, please provide a full cmd log. Then i see where the issue might be,
Also open up a cmd and run
pip cache purge
Then delete the venv folder and relaunch the webui-user.bat
now it starts but it crashes the computer after the last step.... it goes VERY fast but the last step crashes the PC
now it starts but it crashes the computer after the last step.... it goes VERY fast but the last step crashes the PC
WHat model and resolution and other txt2img settings did you used when it crashes? Please provide more information.
now it starts but it crashes the computer after the last step.... it goes VERY fast but the last step crashes the PC
WHat model and resolution and other txt2img settings did you used when it crashes? Please provide more information.
ive used all sizes of resulotions, ffor now the only one that dont really crash hard the PC is the 512*512, ive used stable difusion 1.5 ,openjourney, opendalle, XDSL open journey and some from civic.AI, the crashes are random, what i belive it is the last step that it hangs for a moment and the gpu goes to 100% no mater the steps or size, is there a correct sampling method(for AMD usage)?
venv "J:\Invoke\stable-diffusion-webui-amdgpu\venv\Scripts\Python.exe" WARNING: ZLUDA works best with SD.Next. Please consider migrating to SD.Next. Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)] Version: v1.10.1-amd-2-g395ce8dc Commit hash: 395ce8dc2cb01282d48074a89a5e6cb3da4b59ab Using ZLUDA in J:\Invoke\stable-diffusion-webui-amdgpu\.zluda Skipping onnxruntime installation. You are up to date with the most recent release. no module 'xformers'. Processing without... no module 'xformers'. Processing without... No module 'xformers'. Proceeding without it. J:\Invoke\stable-diffusion-webui-amdgpu\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning:
pytorch_lightning.utilities.distributed.rank_zero_onlyhas been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from
pytorch_lightning.utilitiesinstead. rank_zero_deprecation( Launching Web UI with arguments: --use-zluda --update-check --skip-ort CivitAI Browser+: Aria2 RPC started 找不到llama_cpp模块 Loading weights [e80275529b] from J:\Invoke\stable-diffusion-webui-amdgpu\models\Stable-diffusion\jibMixRealisticXL_v140CrystalClarity.safetensors Creating model from config: J:\Invoke\stable-diffusion-webui-amdgpu\repositories\generative-models\configs\inference\sd_xl_base.yaml J:\Invoke\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py:1150: FutureWarning:
resume_downloadis deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use
force_download=True`.
warnings.warn(
Running on local URL: http://127.0.0.1:7860
To create a public link, set share=True
in launch()
.
Startup time: 88.8s (prepare environment: 154.9s, initialize shared: 2.2s, other imports: 3.4s, load scripts: 1.4s, initialize extra networks: 0.1s, create ui: 1.7s, gradio launch: 0.4s).
Applying attention optimization: Doggettx... done.
Model loaded in 32.4s (load weights from disk: 0.9s, create model: 1.5s, apply weights to model: 28.4s, apply half(): 0.1s, move model to device: 0.2s, load textual inversion embeddings: 0.3s, calculate empty prompt: 0.9s).
Reusing loaded model jibMixRealisticXL_v140CrystalClarity.safetensors [e80275529b] to load openjourney-v4.ckpt [02e37aad9f]
Loading weights [02e37aad9f] from J:\Invoke\stable-diffusion-webui-amdgpu\models\Stable-diffusion\openjourney-v4.ckpt
Creating model from config: J:\Invoke\stable-diffusion-webui-amdgpu\configs\v1-inference.yaml
J:\Invoke\stable-diffusion-webui-amdgpu\venv\lib\site-packages\huggingface_hub\file_download.py:1150: FutureWarning: resume_download
is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use force_download=True
.
warnings.warn(
Applying attention optimization: Doggettx... done.
Model loaded in 2.1s (create model: 0.4s, apply weights to model: 1.4s, calculate empty prompt: 0.2s).
`
That could be a vae related issue. The VAE Step is the last Step. Best is to install the "Tiled DIffusion & Tiled VAE" (Multidiffusion) Extension via the Extensions Tab. https://github.com/pkuliyi2015/multidiffusion-upscaler-for-automatic1111
After the install, restart the webui completely. Then only Enable the Tiled VAE function in txt2img and try again.
Let me know if it worked.
Checklist
What happened?
wont start whatever i do.......
Steps to reproduce the problem
1 download 2 COMMANDLINE_ARGS= --use-directml --medvram --opt-sub-quad-attention --opt-split-attention --no-half-vae --upcast-sampling
What should have happened?
should start and run....
What browsers do you use to access the UI ?
Mozilla Firefox, Google Chrome
Sysinfo
H:\AMD stable\stable-diffusion-webui-amdgpu\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning:
main()
File "H:\AMD stable\stable-diffusion-webui-amdgpu\launch.py", line 29, in main
filename = launch_utils.dump_sysinfo()
File "H:\AMD stable\stable-diffusion-webui-amdgpu\modules\launch_utils.py", line 698, in dump_sysinfo
text = sysinfo.get()
File "H:\AMD stable\stable-diffusion-webui-amdgpu\modules\sysinfo.py", line 46, in get
res = get_dict()
File "H:\AMD stable\stable-diffusion-webui-amdgpu\modules\sysinfo.py", line 119, in get_dict
"Extensions": get_extensions(enabled=True, fallback_disabled_extensions=config.get('disabled_extensions', [])),
AttributeError: 'str' object has no attribute 'get'
pytorch_lightning.utilities.distributed.rank_zero_only
has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it frompytorch_lightning.utilities
instead. rank_zero_deprecation( Traceback (most recent call last): File "H:\AMD stable\stable-diffusion-webui-amdgpu\launch.py", line 48, inConsole logs
Additional information
No response