-
### Describe the bug
Whenever I try to load the model, error shows
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
amd is not suppo…
-
Hello,
Please help to resolve the following issue.
I built my own recipe based on _egs2/librispeech/asr1_.
I was able to successfully run all the stages, where I used GPU for decoding.
However…
-
title, and to be clear, does llama generate eos tokens? because when i increase the max tokens limit it kept on generating the user's questions and stuff too, although in the generator.py i found logi…
-
model: llama-13B-4bit-128g
exllama:
```
(exllama) user@debian:~/AI/exllama$ python test_benchmark_inference.py -d ~/AI/2oobabooga/text-generation-webui/models/llama-13b-4bit-128g/ -p
-- Loadi…
-
when I follow pip install -e ".[gpu]",I find the error about mosaicml-streaming
#-------------------------------------------------------------------------------------------
root@7730f5bd29fa:/hom…
-
when I follow pip install -e ".[gpu]",I find the error about mosaicml-streaming
#-------------------------------------------------------------------------------------------
root@7730f5bd29fa:/hom…
-
Hi,
I have used `--model-control-mode=explicit` option start the triton server without any model loading.
```
mpirun --allow-run-as-root -n 1 /opt/tritonserver/bin/tritonserver --model-contro…
-
Hi I was attempting to use Langchain with a transformers "text-generation" transformers pipeline as described in the video here (this was in the main readme of this repo, so I guess it is somewhat ap…
-
### System Info
# System Info
Running on SageMaker Studio g4dn 2xlarge.
```
!cat /etc/os-release
PRETTY_NAME="Debian GNU/Linux 10 (buster)"
```
```
!transformers-cli env
- `transformers…
-
### Bug Description
I want to reproduce [HuggingFace LLM - StableLM](https://gpt-index.readthedocs.io/en/latest/examples/customization/llms/SimpleIndexDemo-Huggingface_stablelm.html)
`response = q…