Closed DK013 closed 2 months ago
Try setting a smaller context size than the default, e.g. -c 4096
. A lot of models have a 128k context size now, which needs more memory than most consumers have in their PCs. Let me know if that doesn't fix it and I'll reopen.
@jart hey that works. thanks for the clarification.
Contact Details
chakraborty.deep013@gmail.com
What happened?
This may be related to #501 I tried loading Phi-3.5-mini-instruct-Q8_0.gguf failed with memory allocation error.
Here's a log with
--strace
flag just in case: log dumpRelevant System specs: CPU: Ryzen 3600X GPU: Radeon RX 5700 XT RAM: 32GB DDR4 3000Mhz
Version
llamafile v0.8.13
What operating system are you seeing the problem on?
No response
Relevant log output