-
With my NVIDIA 1050TI and Ubuntu 23.04 WSL, the model takes a long time to load—about three minutes—but the response is, I believe, faster than with other tools.
Once the first question was answered,…
-
Steps to reproduce:
Create a model with no template, f16 gguf.
Use github.com/jmorganca/olllama/api Go client.
Set stopwords to various strings that might be emitted, such as other users in the "ch…
-
Running GPT-4 to switch folders is expensive, can a secondary api be integrated for switching folders?
-
### The Feature
Starting this issue to ensure LiteLLM is compatible with OpenAI v1.0.0
## The main goal of this issue:
If a user has OpeAI v1.0.0 their OpenAI calls through litellm should not…
-
This might need to be split into multiple separate Issues. If so please let me know.
## Issues Relating to My Goals (and possibly other people's) with AnythingLLM
There might not be a workaround…
-
### Question Validation
- [X] I have searched both the documentation and discord for an answer.
### Question
Hi,
I really enjoyed reading through the entire llama index documentation to get some…
-
Steps I followed:
- I installed the newest GoLang using winget
- I cloned the repro
- I executed `go build .`
- After initial library download build fails with the error message:
> \# github.c…
-
https://github.com/alingse/go-linter-runner/blob/main/jobs/testifylint.yaml
```
name: testifylint
linter: testifylint
workdir: .
install: go install github.com/Antonboom/testifylint@latest
inc…
-
It seems that mixtral example is giving non-sensical output when used with non-default arguments. Is this expected?
Here is an example
```
python mixtral.py --model_path ../../mixtral/ -m 2000 --pr…
-
I ran into the same bug as https://github.com/karthink/gptel/issues/135
But ultimately even if I didn't run into this bug, I think the transient menu is not as convenient if you have a big list of …