Open axel7083 opened 3 weeks ago
Should be an easy fix: replace MODEL_ENDPOINT to use inference_server_container_ip:8000
I am not able to make a request from a container A to another container B using the ip of the container B I opened https://github.com/containers/podman/issues/24260 to have some background and more information
Following the response from podman, should be fixed in 5.3, keeping this open for now
Bug description
We are using
http://host.containers.internal:<port>
for theMODEL_ENDPOINT
since we removed the inference server of the Pod of the recipe. The solution was introduced in https://github.com/containers/podman-desktop-extension-ai-lab/pull/1503Operating system
Fedora 40
Installation Method
Other
Version
next (development version)
Steps to reproduce
check model availability
Relevant log output
No response
Additional context
cc @jeffmaury