Closed T9es closed 1 year ago
Do you have a way to reproduce it stably?
Not really. This just keeps happening no matter what I do. I have no clue what could be the cause or how to even come close to this problem. All I know that it was working before the recent update. After that it was a no-go. Maybe some VENV got messed up?
DefaultCPUAllocator: not enough memory
This is a problem that has been around for a while, maybe a recent update made it prone to it, but it wouldn't be a new problem. I haven't found out why.
Oh my god, I have a clue to what might have happened. I was messing around with my new SSD yesterday, might be something related to page file. I will try going into this later, since I'm not home as of now and update if that fixed it.
This would explain some things crashing, like discord and chrome, although I have no clue why there would be crashes while 16gigs from 32 are being used.
Seems like this was a page file issue. I'm going to run some tests on my RAM, since that's the main part that could have failed. After setting the page file to the previous settings, everything is working fine.
The problem exist, it's happening to me too. On Colab, from time to time, it collapses because of the excessive use of RAM.
I'm re-opening the issue, as this issue happens from time to time. I've been playing around with generations, and I get DefaultCPUAllocator errors from time to time. RAM usage at max is about 60-70% from 32gb. After the error, I need to restart the AI, since it will not want to generate anything.
I saw people referencing that it's some kind of RAM issue (not enough of it), but as you can see, that's not the case here.
Try Windows' memory test.
Already did. No issues, 10 tests, even switched RAM slots for another set of 10 tests.
My baseline to create image is always 608 x 768, and then use Hires fix 1.8 times, to get approximate Instagram 4:5 resolution. Final Size: 1094x1382, while Instagram 4:5 is 1080x1350. So I need that resolution in one click.
For me since it's 1.3.0 and now 1.3.1, that 608 x 768 resolution and 1.8 time upscale on Hires fix to get my target resolution cannot be used (VRAM always out of memory), and I have to lower my resolution or upscale value, and the result I get is lower on my final resolution. I also trying to setting up Hires steps as half of Sampling steps also doesn't work.
For information, my machine is RTX 3060 Laptop with 6GB VRAM. I always use --xformers --medvram
The only possible way for me to reach my target resolution while preserve detail is using "img2img" with custom setting upscaler mimicking what Hires fix does, and it's successful, there's no VRAM out of memory. It's weird because it's basicaly a same thing. The custom setting is located on Settings: Upscalling > Upscaler for img2img > change it to upscaler that I previously use on Hires fix.
But it's now I have to switch tab to upscale, everytime. It's working but it's not convinience.
And then everyone on another forum and other issue related to this always suggest to use Doggettx as Optimization method after 1.3.0
And I swear I already tried it. I tried to remove xformers and it's only --medvram on my args, and using Doggettx on from Settings, already restart the command prompt, but it's still not avail at all for that resolution and Hires fix value for my target resolution.
It is real, Doggettx is not for every machine, in my case even when it's on, Hires fix still out of memory.
So I'm affraid it's my machine fault or some Windows settings/update/drivers does something bad. So I tried a create a new folder to git clone and hard reset the latest known good A1111, for me v1.2.1 (hash 89f9faa), and install that version from scratch on another folder.
Running from that folder (v1.2.1 89f9faa) the Hires fix can be used as before, with upto 1.8x upscale, but on 1.3.0 and 1.3.1 it's indeed cannot.
So it's not because my system, because it's clear that Hires fix on v1.2.1 89f9faa could work fine.
I'm not complaining, I just pointed out that something is wrong with 1.3.0 and 1.3.1, and Doggettx is not my answer, and I need help.
Closing this down, as it's no longer an issue for me. Has been working fine since the last month or so without any issues. Still related to page file on my end.
Is there an existing issue for this?
What happened?
After updating with git pull, the latest release seems to have A LOT of issues. In my case, the models never seem to leave RAM. When running one instance, I go from 7gb RAM usage, up to 16gb in seconds. I have 32gigs, but this causes A LOT of instability in the system. Can't generate images that I did generate yesterday, or even today morning before updating.
On top of this, here's a weird screenshot from my dedicated SD card. See the issue? There's VRAM available, but it's not getting taken by SD.
Steps to reproduce the problem
There are no steps. Just start the UI and run a regular generation, like per usual. No specific settings, no loras. Just simply typing "yes" into the prompt window and hitting generate.
What should have happened?
Everything gets loaded, RAM gets cleared, there are no issues with VRAM allocation.
Commit where the problem happens
22bcc7be428c94e9408f589966c2040187245d81
What platforms do you use to access the UI ?
Windows 10
What browsers do you use to access the UI ?
Any browser, chrome, firefox, edge, opera.
Command Line Arguments
List of extensions
Don't matter. I run a clean install with no extensions on my second card and the issue was still there.
Console logs
Additional information
2 cards with the same issue. 3060TI (8gb) and a 2060 (6gb). The 2060 is without any extensions, it's vanilla running with --xformers and --api.