Closed iridium-soda closed 2 weeks ago
Also, here is my custom models_config.yaml
:
model_providers:
############################ Local ############################################
# Uncomment this provider if you want to use local models providers #
# using ollama and infinity model server #
###############################################################################
# - provider_name: local-ollama
# api_format: openai
# base_url: http://ollama-server:11434/v1/
# api_key_env_var: ""
# llm_model_ids:
# - "qwen2:1.5b"
# embedding_model_ids: []
# reranking_model_ids: []
# default_headers: {}
# - provider_name: local-infinity
# api_format: openai
# base_url: http://infinity-server:7997/
# api_key_env_var: INFINITY_API_KEY
# llm_model_ids: []
# embedding_model_ids:
# - "mixedbread-ai/mxbai-embed-large-v1"
# reranking_model_ids:
# - "mixedbread-ai/mxbai-rerank-xsmall-v1"
# default_headers: {}
# - provider_name: faster-whisper
# api_format: openai
# base_url: http://faster-whisper:8000
# api_key_env_var: ""
# llm_model_ids: []
# embedding_model_ids: []
# reranking_model_ids: []
# audio_model_ids:
# - "Systran/faster-distil-whisper-large-v3"
# default_headers: {}
############################ OpenAI ###########################################
# Uncomment this provider if you want to use OpenAI as a models provider #
# Remember to set `OPENAI_API_KEY` in container environment #
###############################################################################
# - provider_name: openai
# api_format: openai
# api_key_env_var: OPENAI_API_KEY
# llm_model_ids:
# - "gpt-3.5-turbo"
# - "gpt-4o"
# embedding_model_ids:
# - "text-embedding-3-small"
# - "text-embedding-ada-002"
# reranking_model_ids: []
# default_headers: {}
############################ TrueFoundry ###########################################
# Uncomment this provider if you want to use TrueFoundry as a models provider #
# Remember to set `TFY_API_KEY` in container environment #
####################################################################################
# - provider_name: truefoundry
# api_format: openai
# base_url: https://llm-gateway.truefoundry.com/api/inference/openai
# api_key_env_var: TFY_API_KEY
# llm_model_ids:
# - "openai-main/gpt-4o-mini"
# - "openai-main/gpt-4-turbo"
# - "openai-main/gpt-3-5-turbo"
# embedding_model_ids:
# - "openai-main/text-embedding-3-small"
# - "openai-main/text-embedding-ada-002"
# reranking_model_ids: []
# default_headers: {}
############################ Mixedbread ###########################################
# Uncomment this provider if you want to use Mixedbread as a models provider #
# Remember to set `MIX_API_KEY` in container environment (./compose.env) #
#####################################################################################
- provider_name: Mixedbread
api_format: openai
base_url: https://api.mixedbread.ai
api_key_env_var: MIX_KEY
llm_model_ids: []
embedding_model_ids:
- 'mixedbread-ai/mxbai-embed-large-v1'
reranking_model_ids:
- "mixedbread-ai/mxbai-rerank-large-v1"
default_headers: {}
############################ Azure ###########################################
# Uncomment this provider if you want to use Azure as a models provider #
# Remember to set `AZURE_API_KEY` in container environment #
###############################################################################
- provider_name: Azure
api_format: openai
base_url: AZURE_ENDPOINT
api_key_env_var: AZURE_KEY
llm_model_ids:
- 'gpt35-16k'
embedding_model_ids: []
reranking_model_ids: []
default_headers: {}
Hi I had the same issue, and in my understanding it is necessary to create a truefoundry api key to make it work. Probably it should be specified in the readme.
Hi I had the same issue, and in my understanding it is necessary to create a truefoundry api key to make it work. Probably it should be specified in the readme.
I used to think open-source projects wouldn’t typically require registration, but I understand if this is necessary here. That said, it would be great if this key requirement could be highlighted earlier in the documentation.
Apologies if my earlier message came off as a complaint! Could you kindly point me in the right direction to get the necessary variable? I wasn’t able to find the API key on truefoundry.cloud
. Could it be the Cluster Token from the cluster instead?
Hi,
I am not a developer of this project so I don't want to say things which end up being not true, so let's wait for them to clarify this point about the necessity of using a truefoundry api key. I agree with you that it should be clarified in the readme.
Anyways what I did to make it work is to create an account on truefoundry (you cannot use your personal email, it must be an organization email like your company email.. Not sure if also university email might work) and give a name to the project or the name of your company. This will create an endpoint like https://your-project-name.truefoundry.cloud/
This should be set as TFY_HOST in the compose.env file. To generate an API key you have to click on "Settings" on the left of the truefoundry page after you logged in. If you don't see settings try to do "ctrl+-" to zoom out, because it is below "Integrations". In my default browser settings I couldn't see it, and I realized after a bit that zooming out I could see that button. There you can create a new personal access token that you copy paste in compose.env.
I am on free tier and I never put any credit card, the only thing is that I am not sure if my token will expire or not.
I hope this helps
Hi,
I am not a developer of this project so I don't want to say things which end up being not true, so let's wait for them to clarify this point about the necessity of using a truefoundry api key. I agree with you that it should be clarified in the readme.
Anyways what I did to make it work is to create an account on truefoundry (you cannot use your personal email, it must be an organization email like your company email.. Not sure if also university email might work) and give a name to the project or the name of your company. This will create an endpoint like https://your-project-name.truefoundry.cloud/
This should be set as TFY_HOST in the compose.env file. To generate an API key you have to click on "Settings" on the left of the truefoundry page after you logged in. If you don't see settings try to do "ctrl+-" to zoom out, because it is below "Integrations". In my default browser settings I couldn't see it, and I realized after a bit that zooming out I could see that button. There you can create a new personal access token that you copy paste in compose.env.
I am on free tier and I never put any credit card, the only thing is that I am not sure if my token will expire or not.
I hope this helps
Thank you so much for your kind help! I’m sure it will be really helpful for me 😊.
Hello, This seems to have been introduced recently, looks like we have unintentionally created a hard dependency on TFY_API_KEY
We'll fix this shortly. You should not need to sign up with TrueFoundry, apologies for the inconvenience
Hello @FabrizioDG @iridium-soda,
This issue has been fixed. Please give it a try and let us know in case of issues. Thanks !
I’m trying to use Azure OpenAI as my LLM provider and the Mixedbread Remote API for reranking and embedding. I’m attempting to run the service using Docker Compose, but I’m encountering an issue preventing the service from starting.
I’m unsure whether there’s a configuration mistake on my end, or if the issue is related to environment variables. Specifically, I’m wondering if the TFY_API_KEY must be explicitly assigned for the service to run correctly.
Questions:
Are there any specific configurations or common pitfalls when integrating Azure OpenAI with Mixedbread API through Docker Compose? Is the
TFY_API_KEY
required for the deployment of the service? Could there be any other setup steps that I might have overlooked? Thank you for your assistance!Here is my error message when trying to deploy: