Open thams opened 3 months ago
@thams try it with
"endpoints": [
{
"type": "tgi",
"url": "https://1212121212.us-east-1.aws.endpoints.huggingface.cloud"
}
]
Also not sure if your template works, at least llama does not seem to have the <|system|>
token. This works for me:
"chatPromptTemplate" : "<|begin_of_text|>{{#each messages}}{{#ifUser}}<|start_header_id|>user<|end_header_id|>\n{{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}}{{/ifUser}}{{#ifAssistant}}<|start_header_id|>assistant<|end_header_id|>\n{{content}}{{/ifAssistant}}{{/each}}",
there seems to be no configuration for .env.local that I can get to work to connect to a Llama3 inference endpoint hosted by HuggingFace cloud (and I can find no examples).
This flavor seems to need a value for endpoint type, one of:
Invalid discriminator value. Expected 'anthropic' | 'anthropic-vertex' | 'aws' | 'openai' | 'tgi' | 'llamacpp' | 'ollama' | 'vertex' | 'genai' | 'cloudflare' | 'cohere' | 'langserve'"
but none of them work.