Closed dhandhalyabhavik closed 2 months ago
Runing on Xeon, I can reproduce the same issue. (And ollama service uses the same port 9000 with tgi service.)
Assign to Lvliang to help fix. Thanks.
already fixed by PR https://github.com/opea-project/GenAIComps/pull/248 @dhandhalyabhavik @feng-intel
Hey @lvliang-intel, Thanks for fixing the issue. I am able to run llama-3.
but not able to run other models because llama-3 is hardcoded at line 24 here
Can you please enable it as REST argument similar to how ollama has?
like this
curl http://localhost:11434/api/generate -d '{
"model": "llama3",
"prompt":"Why is the sky blue?"
}'
HI @lvliang-intel Do we need to modify this hardcode here
@dhandhalyabhavik You can try to update "llama3" to other model.
Then I will have to modify docker images every time I want to try new model, it should be just REST argument as provided above.
@dhandhalyabhavik, PR https://github.com/opea-project/GenAIComps/pull/280 fixed this issue.
Thank you. Closing this now.
After building ollama image,
When tried running it, it failed with following error,