Please provide a detailed written description of what you were trying to do, and what you expected llama.cpp to do as an enhancement.
Ollama uses a compiled version of llama.cpp . Letting end-uses re-compile ollama and llama.cpp in order to enable the usage of integrated GPUs is problematic. I would enjoy LLAMA_HIP_UMA to be a dynamic setting that can be enabled regardless of the compile time flags.
I think right now there are three ways to get iGPUs working in ollama:
Feature Description
Please provide a detailed written description of what you were trying to do, and what you expected
llama.cpp
to do as an enhancement.Ollama uses a compiled version of llama.cpp . Letting end-uses re-compile ollama and llama.cpp in order to enable the usage of integrated GPUs is problematic. I would enjoy LLAMA_HIP_UMA to be a dynamic setting that can be enabled regardless of the compile time flags.
I think right now there are three ways to get iGPUs working in ollama:
See also https://github.com/ollama/ollama/issues/2637
Possible Implementation
?