Closed louie-tsai closed 1 week ago
P3-Medium
Ubuntu
Xeon-SPR
Single Node
1.0
The ChatQnA REAMDE TGI instruction below is not aligned with our LLM gen component. https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA/docker_compose/intel/hpu/gaudi
LLM GenAI component implementation: https://github.com/opea-project/GenAIComps/blob/main/comps/llms/text-generation/tgi/llm.py#L41
just follow the README. it should still work for testing since TGI also has generate endpoint, but our microservice use v1/completion instead.
No response
fixed with commit : https://github.com/intel-ai-tce/GenAIExamples/commit/aa314f6757c503f0b10a6481b1e2eb00db13fa6e
Priority
P3-Medium
OS type
Ubuntu
Hardware type
Xeon-SPR
Installation method
Deploy method
Running nodes
Single Node
What's the version?
1.0
Description
The ChatQnA REAMDE TGI instruction below is not aligned with our LLM gen component. https://github.com/opea-project/GenAIExamples/tree/main/ChatQnA/docker_compose/intel/hpu/gaudi
LLM GenAI component implementation: https://github.com/opea-project/GenAIComps/blob/main/comps/llms/text-generation/tgi/llm.py#L41
Reproduce steps
just follow the README. it should still work for testing since TGI also has generate endpoint, but our microservice use v1/completion instead.
Raw log
No response