-
This issue is used to document the LLM, embedding, reranker, etc. models that need to be integrated with RAGFlow.
- [x] Azure Open AI service
- [x] Google Gemini #1036
- [x] Mixtral AI #433
- […
-
I believe, in order to resolve https://github.com/mudler/LocalAI/pull/1446, go-llama.cpp needs to be built against at least version 799a1cb13b0b1b560ab0ceff485caed68faa8f1f of llama.cpp to enable mixt…
-
# Description
At the moment, all current prompts for personas are written/tuned for Llama 3. However, with the introduction of Integrations and the subsequent move of various parts of the ML stack to…
-
### Feature request
### System Info
transformers version: 4.36.2
### Who can help?
don't have a clue about this
### Information
Refer to llama2 modeling code, I want to add attention bias …
-
### Your current environment
Edit 1
```text
Collecting environment information...
PyTorch version: 2.2.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorc…
-
groq has jawdroppingly fast access to mixtral. Currently you can use the UI and API for no cost. There is throttling but it seems quite generous
it's easy to use via the awesome litellm
See http…
-
### Checklist
- [X] 1. I have searched related issues but cannot get the expected help.
- [X] 2. The bug has not been fixed in the latest version.
- [X] 3. Please note that if the bug-related iss…
-
So I followed the [instructions](https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/GPU/HF-Transformers-AutoModels/Model/mixtral) to the best of my understanding and got:
`(llm) c:…
-
MIT emails and dormspam event data is classified as [medium risk](https://infoprotect.mit.edu/risk-classifications#quicktabs-data_risk_levels=1) information, and sending this information to proprietar…
-
chat with multiple model available with @
for example:
1. setup : above chat box rather than pick 1 model able to pick group model (group model set in the setting)
2. use: type prompt in prompt bo…