-
**Is your feature request related to a problem? Please describe.**
I want to implement a feedback sending mechanism from the front-end of a chat bot to a **python** backend.
For LangChain, it is pos…
-
### Here's example of gradio chat, where user give a promt
with gr.Blocks() as demo:
chatbot = gr.Chatbot(value=[[None, "Hello, How can you assist you today?"]])
msg = gr.Textbox()
c…
-
### Initial Checks
- [X] I have searched GitHub for a duplicate issue and I'm sure this is something new
- [X] I have read and followed [the docs & demos](https://github.com/modelscope/modelscope-age…
-
Cc: @adamcogan @jimmidier @duanxinhuan
To myself,
### Pain
Feedback from @adamcogan
> In that current Sprint there is something else to think about than just bugs. We should have a PBI that …
-
Input:
```json
{
"config_id": "config",
"messages": [
{
"role": "user",
"content": "How to cook egg\n"
}
]
}
```
Logs:
```
Event Utt…
-
1. **Motivation** - We could send the current conversation path to the llm and with the answer provided by the bot model to generate a personalized response for the user. Every user could have a perso…
-
Might make it easier to find?
-
I am using VSCode WSL2, Ubuntu 22.04 and Docker Engine v24.0.6
The .env file contains:
LLM=mistral #or any llama2:7b Ollama model tag, gpt-4, gpt-3.5, or claudev2
EMBEDDING_MODEL=sentence_tran…
-
I'm trying to get paper-qa working locally, but the documentation doesn't properly show how that should be done. I'm working from the documentation for [local usage](https://github.com/Future-House/pa…
-
AS a Bürokratt end user
I WANT TO receive information or a hint from the Bureaucrat when I am waiting for a response from the bot, whether it generates a response or is broken
SO THAT I can understa…