-
**Why**
Users will be able to leverage Fireworks AI's language models within big-AGI, expanding the range of AI capabilities.
**Description**
Integrate Fireworks AI as a new language model provid…
-
### Describe the bug
I get some sort of LiteLLM error when I try to connect to the openai api in oogabooga.
It works fine to connect to the API from SillyTavern.
On a tangent: It looks like LiteL…
-
### Your current environment
VLLM image: v0.5.4
hardware: RTX4090
gpu driver: 550.78
model: qwen1.5-14b-chat-awq
launch cmd: enable-prefix-caching
### 🐛 Describe the bug
```
2024-08-30T15:30…
-
**Summary**
This enhancement aims to separate the usage of local large language models (LLMs) from LLMs accessed through APIs, and reduce the dependency on installed packages like NVIDIA and PyTorch.…
-
This is a very popular method of interacting with models. Frequent updates and good performance. Can it be used with your program?
-
More of a question than an issue, but I wanted to see if you have a recommendation for creating audit trails or plans to support it more directly from a library interface.
I would like to have tool…
-
When training the Qlora+unsloth with SFT qwen2-72B-Instruct model, an error occurs with the message "FlashAttention only supports fp16 and bf16 data types." Below is the specific error traceback:
[ra…
-
### Bug Description
Imagine an instantiated Vertex ai model (LLM). I'm trying to access the "achat" interface to chat with the gemini model (gemini-1.5-flash-002)
And I'm getting the following e…
-
Hello,
When using `BedrockChat` through langchain, the streaming functionality does not work. `Claude3` models requires `BedrockChat` interface to be used. When I switch to the `Bedrock` interface …
-
- [ ] [system-2-research/README.md at main · open-thought/system-2-research](https://github.com/open-thought/system-2-research/blob/main/README.md?plain=1)
# OpenThought - System 2 Research Links
He…