-
### Describe the bug
Calling `InferenceClient.conversational` according to the [docs](https://huggingface.co/docs/inference-endpoints/main/en/supported_tasks#conversational) results in a `400 Client …
-
For debugging or other traceability purposes it is sometimes useful to see the final prompt text as sent to the completion model.
It would be good to have a mechanism that logged or otherwise sur…
-
### Checked other resources
- [X] I searched the Codefuse documentation with the integrated search.
- [X] I used the GitHub search to find a similar question and didn't find it.
- [X] I am sure that …
-
**问题描述 / Problem Description**
bge-m3不兼容 / 显存异常分配
和issue #4101 相似
**复现问题的步骤 / Steps to Reproduce**
1. 配置model_config.py 制定模型为bge-m3
2. 运行后,请求接口,根据content重建知识库:/knowledge_base/recreate_vector_st…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
Hi: I'm trying to swap out:
`llm = ChatOpenAI(temperature = temperature, model_name = 'gpt-3.5-turbo', max_tokens = max_output_size)
`
with:
`llm = HuggingFaceHub(repo_id = 'mosaicml/mpt-7b-…
-
I attempted to run the chatbot-rag-app using an Azure OpenAI model, and found that when I entered a question in the prompt box in the GUI, the service would never produce a response.
I was able to …
-
### Describe the bug
I'm developing a chatbot based on the RAG-QA.
I'm getting an Error on the Gradio front-end even though the backend works properly.
I'm uploading 1 page from Wikipedia using…
-
Originally posted by **0xSynth** January 17, 2024
Hi, I can't run ollama mixtral model on my desktop, however I have a server to which I would like to connect (via ngrok), however for some reason I c…
-
### Describe the bug
Hi, @dosu-bot
I am using langchain agent and Streamlit to create a chatbot who can analyse user's question, generate a SQL query, execute SQL query, and translate the results …