Closed noobjam closed 3 months ago
Hi @noobjam, sorry for the inconvenience. Model mistralai/Mixtral-8x7B-Instruct-v0.1
is not available on our Inference API. Agree the error message should be more explicit than the current HTTP 504 you are getting. We are working on fixing this. In the meantime, I advice you to use either https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct or https://huggingface.co/meta-llama/Meta-Llama-3.1-70B-Instruct for your use case. Otherwise you can deploy mistralai/Mixtral-8x7B-Instruct-v0.1
as a dedicated Inference Endpoints but that would require you to pay for the server.
Describe the bug
I encountered a 504 Gateway Timeout error when attempting to use the mistralai/Mixtral-8x7B-Instruct-v0.1 model via the Hugging Face Inference API. The error occurred during model invocation and prevented successful completion of the request.
Reproduction
`from langchain.llms import HuggingFaceEndpoint from langchain.prompts import PromptTemplate from langchain.chains import LLMChain
Initialize the HuggingFaceEndpoint
llm = HuggingFaceEndpoint( repo_id="mistralai/Mixtral-8x7B-Instruct-v0.1", max_new_tokens=300, top_k=50, top_p=0.99,
temperature=0.01,
huggingfacehub_api_token="xxxxxxxxxxxxxxxxxxxxxxx" # API token )
prompt = PromptTemplate(template="Translate the following text to French: {text}", input_variables=["text"])
chain = LLMChain(llm=llm, prompt=prompt)
result = chain.run({"text": "Hello, how are you?"})
print(result)`
Logs
No response
System info