Closed igordcard closed 6 days ago
MODELNAME="meta-llama/Meta-Llama-3-8B" meta-llama need special permission with HUGGINGFACE TOKEN, previously we didn't pass the TOKEN to TGI inference server and you can use models like intel/xxx. The Reorg of helm charts added support for passing HF_TOKEN to tgi server, you can try after that merged (Pay attention to the readme file change with helm dependency update) Before that, you can try with other models like Intel/xxx instead of the meta-llama
Thanks Dolpher, your fix is working for me.
I see the following in
chatqna-tgi
's logs (in K8s), installed via https://github.com/opea-project/GenAIInfra/tree/main/helm-charts/chatqna even after passing my correct hugging face token and accepting the conditions to accessing the gated repo for llama3:kubectl logs chatqna-tgi-f455cdb9b-bqwfv
: