Open fairyqb opened 5 months ago
Hello fairyqb,
Thank you for bringing this to our attention. It seems like you are experiencing high CPU usage with llama-cpp on your server. We understand this can be a concern and would like to help. Please provide us with more information regarding your LocalAI configuration and any commands used to run llama-cpp. It would also be helpful to get the output of the server load and the top
Do you have a gpu? If so, I’m seeing the same issue, on docker, at least. https://github.com/mudler/LocalAI/discussions/2191
llama-cpp CPU 1500%,Very slow...
my server:centos,20 core, 32GB memory