Open bard-regard opened 1 year ago
492 works!
N.B *all run at 1.3it/s
I wonder if there is any release between the two which has this performance boost and also works?#492 runs 1.22it/s and #511 runs 3-4it/s
Being quite new to all of this I tried setting 492 to use the 784x784 resolution and adjusted steps to 80 & cfg scale to 8.8. The exe started downloading extra files, I think for the resolution. It seems to be stuck for at least 20m after this line. Maybe larger images do take much longer to setup env; I'm not sure. Looking at the releases, I might try 506 next as thats a nice bug fix. For now, I'm coming off.
...
torch\storage.py:899: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage()
storage = cls(wrap_storage=untyped_storage)
torch\fx\node.py:244: UserWarning: Trying to prepend a node to itself. This behavior has no effect on the graph.
warnings.warn("Trying to prepend a node to itself. This behavior has no effect on the graph.")
No vmfb found. Compiling and saving to D:\stable-diffusion\shark\492\unet1_64_784_784_fp16_stabilityai_stable_diffusion_2_1_vulkan-00000000-0300-0000-0000-000000000000.vmfb
Using target triple -iree-vulkan-target-triple=rdna2-unknown-windows from command line args
v.506
Also, the only working scheduler is SharkEulerDiscrete
, everything else produces this brown picture with a mild texture
N.B Looks like this might be related to a .yaml
file, mentioned on reddit
System:
UI / Prompt
stdout - Click me
``` No vmfb found. Compiling and saving to D:\stable-diffusion\shark\clip64_512_512_fp16_stabilityai_stable_diffusion_2_1_base_vulkan-00000000-0300-0000-0000-000000000000.vmfb Using target triple -iree-vulkan-target-triple=rdna2-unknown-windows from command line args Saved vmfb in D:\stable-diffusion\shark\clip64_512_512_fp16_stabilityai_stable_diffusion_2_1_base_vulkan-00000000-0300-0000-0000-000000000000.vmfb. Downloading (…)tokenizer/vocab.json: 100%|████████████████████████████████████████| 1.06M/1.06M [00:00<00:00, 1.70MB/s] huggingface_hub\file_download.py:129: UserWarning: `huggingface_hub` cache-system uses symlinks by default to efficiently store duplicated files but your machine does not support them in C:\Users\BardzyBear\.cache\huggingface\hub. Caching files will still work but in a degraded version that might require more space on your disk. This warning can be disabled by setting the `HF_HUB_DISABLE_SYMLINKS_WARNING` environment variable. For more details, see https://huggingface.co/docs/huggingface_hub/how-to-cache#limitations. To support symlinks on Windows, you either need to activate Developer Mode or to run Python as an administrator. In order to see activate developer mode, see this article: https://docs.microsoft.com/en-us/windows/apps/get-started/enable-your-device-for-development Downloading (…)tokenizer/merges.txt: 100%|██████████████████████████████████████████| 525k/525k [00:00<00:00, 1.32MB/s] Downloading (…)cial_tokens_map.json: 100%|█████████████████████████████████████████████| 460/460 [00:00<00:00, 465kB/s] Downloading (…)okenizer_config.json: 100%|█████████████████████████████████████████████| 824/824 [00:00<00:00, 832kB/s] Downloading artifacts for model unet64_512_512_fp16_stabilityai_stable_diffusion_2_1_base... 100%|█████████████████████████████████████████████████████████████████████████████████| 156/156 [00:00<00:00, 1.12kB/s] 100%|██████████████████████████████████████████████████████████████████████████████| 64.3k/64.3k [00:00<00:00, 193kB/s] 100%|█████████████████████████████████████████████████████████████████████████████████| 640/640 [00:00<00:00, 51.0kB/s] 100%|████████████████████████████████████████████████████████████████████████████████| 289k/289k [00:00<00:00, 544kB/s] 100%|█████████████████████████████████████████████████████████████████████████████| 1.61G/1.61G [00:51<00:00, 33.8MB/s] No vmfb found. Compiling and saving to D:\stable-diffusion\shark\unet64_512_512_fp16_stabilityai_stable_diffusion_2_1_base_vulkan-00000000-0300-0000-0000-000000000000.vmfb Using target triple -iree-vulkan-target-triple=rdna2-unknown-windows from command line args Saved vmfb in D:\stable-diffusion\shark\unet64_512_512_fp16_stabilityai_stable_diffusion_2_1_base_vulkan-00000000-0300-0000-0000-000000000000.vmfb. 50it [00:16, 3.00it/s] 0%| | 0/1 [00:03, ?it/s] Traceback (most recent call last): File "gradio\routes.py", line 374, in run_predict File "gradio\blocks.py", line 1017, in process_api File "gradio\blocks.py", line 835, in call_function File "anyio\to_thread.py", line 31, in run_sync File "anyio\_backends\_asyncio.py", line 937, in run_sync_in_worker_thread File "anyio\_backends\_asyncio.py", line 867, in run File "apps\stable_diffusion\scripts\txt2img.py", line 241, in txt2img_inf File "apps\stable_diffusion\src\pipelines\pipeline_shark_stable_diffusion_txt2img.py", line 128, in generate_images File "apps\stable_diffusion\src\pipelines\pipeline_shark_stable_diffusion_utils.py", line 93, in decode_latents File "shark\shark_inference.py", line 138, in __call__ File "shark\shark_runner.py", line 93, in run File "shark\iree_utils\compile_utils.py", line 381, in get_results File "iree\runtime\function.py", line 130, in __call__ File "iree\runtime\function.py", line 154, in _invoke RuntimeError: Error invoking function: D:\a\SHARK-Runtime\SHARK-Runtime\c\runtime\src\iree\hal\drivers\vulkan\native_semaphore.cc:155: RESOURCE_EXHAUSTED; overflowed timeline semaphore max value; while invoking native function hal.fence.await; while calling import; [ 1] native hal.fence.await:0 - [ 0] bytecode module.forward:35180 [