Closed Jay19751103 closed 7 months ago
Hi lshqqytiger
Is it possible without reload for every batch inference ?
Yes. But it will use more vram. I'll refactor current codebase of this repository based on my new implementation. Before that, please consider using olive branch of vladmandic/automatic which is much developed.
Checklist
What happened?
When use directml with olive optimized model Just change batch_count from 1 to 30 Everytime it will reload the model, the speed by using olive will around 41s but using this webui will take around 128 to 129 seconds with batch count 30 (I tried 3080TI or 7900XTX, it's same behavior) The reason is that every time it will reload the model. Could you change it as olive ? Olive use a loop the run inference only by num_image parameter
Steps to reproduce the problem
What should have happened?
Inference time should be shorter to improve batch count setting
What browsers do you use to access the UI ?
No response
Sysinfo
GPU card : 3080TI or 7900XTX (VRAM 24G) CPU : AMD 7700X System memory : 32G
Console logs
Additional information
No response