-
## Summary
Support any OpenAI compatible endpoints, such as tabbyAPI, vLLM, ollama, etc.
I am running Qwen2.5-coder 32B with [tabbyAPI](https://github.com/theroyallab/tabbyAPI) which is a OpenAI …
-
### 🚀 The feature, motivation and pitch
I would like to propose a feature that creates a gateway for a model, allowing for multiple models to generate content within a specific context. This is simil…
-
As discussed in our previous meeting, in our ML model interface we should distinguish between parameters used for training and evaluation.
- Allows for different configurations being used in the re…
-
**Is your feature request related to a problem?**
to support conversational search, when sending the request to the remote model, we not only needs to send the questions, but also the historical cont…
-
### Issue type
Bug
### Have you reproduced the bug with TensorFlow Nightly?
Yes
### Source
source
### TensorFlow version
tf 2.14.0
### Custom code
Yes
### OS platform and…
-
Back in the day, the evaluator for a `@model` would look like
```julia
function demo(rng, model, varinfo, context, sampler)
...
end
```
or something like this.
But when we started making us…
-
### Please update gptel first -- errors are often fixed by the time they're reported.
- [X] I have updated gptel to the latest commit and tested that the issue still exists
### Bug Description
setq…
-
### Description
I just migrated our solution to .NET 9 and now im getting a Binding Failure in a DataTemplate used in a ListView.
Error message: Mismatch between the specified x:DataType (MY_MODEL) a…
-
Right now, gptel lets you add files and buffers to the context. But if I understand right, from from the code of functions like `gptel-context--insert-buffer-string` and `gptel-context--file-string`, …
algal updated
3 weeks ago
-
### What happened?
1. llama.cpp failed with Vulkan-supported and quantized model in Android Termux .
2. It run well in CPU mode with quantized model and fp16 model.
3. but if gpu layer is set non-…