Closed Spycsh closed 1 week ago
Seems the doc is simply the fork of TEI so the README does not fully apply to TEI Gaudi.
@Spycsh you are right, configurations that are officially supported in this fork are pointed out in main README
@Spycsh @kdamaszk Thanks for bringing up this issue. We have also experienced this as a blocker, and we have a need to run llava in TGI-Gaudi. Right now we have TGI-Gaudi running various LLMs in our product, but we have to use a workaround for llava, as we hit this issue. Given that Optimum Habana now officially supports llava_next, it would be great to get this in this TGI fork. See, for example: https://github.com/search?q=repo%3Ahuggingface%2Foptimum-habana+llava&type=pullrequests
Is there any active work ongoing to support llava in this TGI fork? If so, great. If not, maybe I could drum up some interest in a contribution (either within our company or with our partners).
Ah, I'm just seeing this: https://github.com/huggingface/tgi-gaudi/pull/193. Not sure how I missed that.
Llava-next support was added in https://github.com/huggingface/tgi-gaudi/pull/187, closing this issue.
System Info
Use the ghcr.io/huggingface/tgi-gaudi:2.0.1 official docker image.
As shown in https://github.com/huggingface/tgi-gaudi/blob/habana-main/docs/source/supported_models.md?plain=1, llava-hf/llava-v1.6-mistral-7b-hf should be supported but it appears not to be supported.
Information
Tasks
Reproduction
Both of following tries fail with the same "Model unsupported" error above
Expected behavior
Since the README claims that llava_next is supported, users should be able to use that model