Closed lstocchi closed 2 months ago
The one from ai-lab-recipes uses library krunkit specific ?
It is primarily intended to be used on macOS with podman and krunkit but Sergio also said that it can be used on Linux with a VMM exposing a Venus-capable virtio-gpu.
The one from ai-lab-recipes uses library krunkit specific ?
It is primarily intended to be used on macOS with podman and krunkit but Sergio also said that it can be used on Linux with a VMM exposing a Venus-capable virtio-gpu.
Did you check that performances with this images are the same compared to the benchmarks we did on MacOS/krunkit ?
Did you check that performances with this images are the same compared to the benchmarks we did on MacOS/krunkit ?
What do you mean? Didn't we use the image generated by the ai-lab-recipes team for the benchmark? I just copy/paste the containerfile so the result should be the same, no?
Did you check that performances with this images are the same compared to the benchmarks we did on MacOS/krunkit ?
What do you mean? Didn't we use the image generated by the ai-lab-recipes team for the benchmark? I just copy/paste the containerfile so the result should be the same, no?
Does not seem to me as the ai-lab-recipe one contains libkrun specific libraries: see https://github.com/containers/ai-lab-recipes/blob/624155c699b7c2abc3c5878fba21ad848b23af25/model_servers/llamacpp_python/vulkan/arm64/Containerfile#L6
Does not seem to me as the ai-lab-recipe one contains libkrun specific libraries: see https://github.com/containers/ai-lab-recipes/blob/624155c699b7c2abc3c5878fba21ad848b23af25/model_servers/llamacpp_python/vulkan/arm64/Containerfile#L6
What does this PR do?
libkrun provider has been added to podman and soon we should have a way to start an inference server that is able to use gpu on a mac. This PR adds the creation of vulkan images so that, eventually, we can use them to start an inference server with gpu capabilities.
Screenshot / video of UI
N/A
What issues does this PR fix or reference?
N/A
How to test this PR?