-
Currently have an LLM engine built on TensorRT-LLM. Trying to evaluate different setups and gains on types.
Was trying to deploy the llama model on a multi-gpu, whereby between the 4 GPUs, I would hav…
-
hello, can i once deploy several models for server ?
whcjb updated
2 months ago
-
Hello, I have a trained model with two branches and two different outputs as well. I tried to change related codes (including configure.conf, cpp_body, fortran_extern_interface, fortran_interface) to …
-
See: https://github.com/himself65/bunchee-issue-593
```shell
> @llamaindex/env@0.1.16 build
> bunchee
Exports File Size
. …
-
**Describe the bug**
I'm using a multi-quantile forecaster on multivariate target data. E.g, a `CatBoostModel(likelihood='quantile', quantile=[0.01, 0.05, 0.1, 0.25, 0.5, 0.75, 0.9, 0.95, 0.99], ..…
-
**🚀 Feature Description**
This is a request for improving the documentation. On the readme, you have a
- List the available speakers and choose a among them:
`$ tts --model_name "//" --list_…
surak updated
1 month ago
-
Does the code support training on multiple datasets? Does your paper involve an experiment where O365, GoldG, and CC1M data are trained together? Do you have this training script? Also, may I ask if t…
-
## Motivation
### Background
To provide more control over the model inputs, we currently define two methods for multi-modal models in vLLM:
- The **input processor** is called inside `LLMEngi…
-
**Submitting author:** @trulsf (Truls Flatberg)
**Repository:** https://github.com/sintefore/TimeStruct.jl
**Branch with paper.md** (empty if default branch): joss_submission
**Version:** 0.9.1
**Edit…
-
### What is the issue?
Currently Ollama can [import GGUF files](https://github.com/ollama/ollama/blob/main/docs/import.md). However, larger models are sometimes split into separate files. Ollama shou…